code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class _lowerCAmelCase ( lowercase ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = ["image_processor"] __UpperCAmelCase : Optional[Any] = "SamImageProcessor" def __init__( self : Any, UpperCAmelCase__ : Dict ): super().__init__(UpperCAmelCase__ ) __lowercase = self.image_processor __lowercase = -1_0 __lowercase = self.image_processor.size["longest_edge"] def __call__( self : Dict, UpperCAmelCase__ : int=None, UpperCAmelCase__ : List[str]=None, UpperCAmelCase__ : List[str]=None, UpperCAmelCase__ : Union[str, Any]=None, UpperCAmelCase__ : Optional[Union[str, TensorType]] = None, **UpperCAmelCase__ : Tuple, ): __lowercase = self.image_processor( UpperCAmelCase__, return_tensors=UpperCAmelCase__, **UpperCAmelCase__, ) # pop arguments that are not used in the foward but used nevertheless __lowercase = encoding_image_processor["original_sizes"] if hasattr(UpperCAmelCase__, "numpy" ): # Checks if Torch or TF tensor __lowercase = original_sizes.numpy() __lowercase ,__lowercase ,__lowercase = self._check_and_preprocess_points( input_points=UpperCAmelCase__, input_labels=UpperCAmelCase__, input_boxes=UpperCAmelCase__, ) __lowercase = self._normalize_and_convert( UpperCAmelCase__, UpperCAmelCase__, input_points=UpperCAmelCase__, input_labels=UpperCAmelCase__, input_boxes=UpperCAmelCase__, return_tensors=UpperCAmelCase__, ) return encoding_image_processor def _lowercase ( self : Union[str, Any], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any]=None, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : Optional[Any]=None, UpperCAmelCase__ : int="pt", ): if input_points is not None: if len(UpperCAmelCase__ ) != len(UpperCAmelCase__ ): __lowercase = [ self._normalize_coordinates(self.target_size, UpperCAmelCase__, original_sizes[0] ) for point in input_points ] else: __lowercase = [ self._normalize_coordinates(self.target_size, UpperCAmelCase__, UpperCAmelCase__ ) for point, original_size in zip(UpperCAmelCase__, UpperCAmelCase__ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: __lowercase ,__lowercase = self._pad_points_and_labels(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = np.array(UpperCAmelCase__ ) if input_labels is not None: __lowercase = np.array(UpperCAmelCase__ ) if input_boxes is not None: if len(UpperCAmelCase__ ) != len(UpperCAmelCase__ ): __lowercase = [ self._normalize_coordinates(self.target_size, UpperCAmelCase__, original_sizes[0], is_bounding_box=UpperCAmelCase__ ) for box in input_boxes ] else: __lowercase = [ self._normalize_coordinates(self.target_size, UpperCAmelCase__, UpperCAmelCase__, is_bounding_box=UpperCAmelCase__ ) for box, original_size in zip(UpperCAmelCase__, UpperCAmelCase__ ) ] __lowercase = np.array(UpperCAmelCase__ ) if input_boxes is not None: if return_tensors == "pt": __lowercase = torch.from_numpy(UpperCAmelCase__ ) # boxes batch size of 1 by default __lowercase = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": __lowercase = tf.convert_to_tensor(UpperCAmelCase__ ) # boxes batch size of 1 by default __lowercase = tf.expand_dims(UpperCAmelCase__, 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": __lowercase = torch.from_numpy(UpperCAmelCase__ ) # point batch size of 1 by default __lowercase = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": __lowercase = tf.convert_to_tensor(UpperCAmelCase__ ) # point batch size of 1 by default __lowercase = tf.expand_dims(UpperCAmelCase__, 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": __lowercase = torch.from_numpy(UpperCAmelCase__ ) # point batch size of 1 by default __lowercase = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": __lowercase = tf.convert_to_tensor(UpperCAmelCase__ ) # point batch size of 1 by default __lowercase = tf.expand_dims(UpperCAmelCase__, 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def _lowercase ( self : Tuple, UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[Any] ): __lowercase = max([point.shape[0] for point in input_points] ) __lowercase = [] for i, point in enumerate(UpperCAmelCase__ ): if point.shape[0] != expected_nb_points: __lowercase = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value], axis=0 ) __lowercase = np.append(input_labels[i], [self.point_pad_value] ) processed_input_points.append(UpperCAmelCase__ ) __lowercase = processed_input_points return input_points, input_labels def _lowercase ( self : List[Any], UpperCAmelCase__ : int, UpperCAmelCase__ : np.ndarray, UpperCAmelCase__ : Any, UpperCAmelCase__ : Tuple=False ): __lowercase ,__lowercase = original_size __lowercase ,__lowercase = self.image_processor._get_preprocess_shape(UpperCAmelCase__, longest_edge=UpperCAmelCase__ ) __lowercase = deepcopy(UpperCAmelCase__ ).astype(UpperCAmelCase__ ) if is_bounding_box: __lowercase = coords.reshape(-1, 2, 2 ) __lowercase = coords[..., 0] * (new_w / old_w) __lowercase = coords[..., 1] * (new_h / old_h) if is_bounding_box: __lowercase = coords.reshape(-1, 4 ) return coords def _lowercase ( self : Tuple, UpperCAmelCase__ : List[Any]=None, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : List[Any]=None, ): if input_points is not None: if hasattr(UpperCAmelCase__, "numpy" ): # Checks for TF or Torch tensor __lowercase = input_points.numpy().tolist() if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ) or not isinstance(input_points[0], UpperCAmelCase__ ): raise ValueError("Input points must be a list of list of floating points." ) __lowercase = [np.array(UpperCAmelCase__ ) for input_point in input_points] else: __lowercase = None if input_labels is not None: if hasattr(UpperCAmelCase__, "numpy" ): __lowercase = input_labels.numpy().tolist() if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ) or not isinstance(input_labels[0], UpperCAmelCase__ ): raise ValueError("Input labels must be a list of list integers." ) __lowercase = [np.array(UpperCAmelCase__ ) for label in input_labels] else: __lowercase = None if input_boxes is not None: if hasattr(UpperCAmelCase__, "numpy" ): __lowercase = input_boxes.numpy().tolist() if ( not isinstance(UpperCAmelCase__, UpperCAmelCase__ ) or not isinstance(input_boxes[0], UpperCAmelCase__ ) or not isinstance(input_boxes[0][0], UpperCAmelCase__ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) __lowercase = [np.array(UpperCAmelCase__ ).astype(np.floataa ) for box in input_boxes] else: __lowercase = None return input_points, input_labels, input_boxes @property def _lowercase ( self : Optional[Any] ): __lowercase = self.image_processor.model_input_names return list(dict.fromkeys(UpperCAmelCase__ ) ) def _lowercase ( self : Optional[int], *UpperCAmelCase__ : List[str], **UpperCAmelCase__ : Union[str, Any] ): return self.image_processor.post_process_masks(*UpperCAmelCase__, **UpperCAmelCase__ )
17
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """facebook/xmod-base""": """https://huggingface.co/facebook/xmod-base/resolve/main/config.json""", """facebook/xmod-large-prenorm""": """https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json""", """facebook/xmod-base-13-125k""": """https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json""", """facebook/xmod-base-30-125k""": """https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json""", """facebook/xmod-base-30-195k""": """https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json""", """facebook/xmod-base-60-125k""": """https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json""", """facebook/xmod-base-60-265k""": """https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json""", """facebook/xmod-base-75-125k""": """https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json""", """facebook/xmod-base-75-269k""": """https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json""", } class lowercase__ ( _UpperCAmelCase ): A__ : int ="""xmod""" def __init__( self : Optional[Any] , UpperCAmelCase_ : Optional[int]=30522 , UpperCAmelCase_ : Union[str, Any]=768 , UpperCAmelCase_ : List[str]=12 , UpperCAmelCase_ : List[str]=12 , UpperCAmelCase_ : List[str]=3072 , UpperCAmelCase_ : Dict="gelu" , UpperCAmelCase_ : str=0.1 , UpperCAmelCase_ : Dict=0.1 , UpperCAmelCase_ : Optional[Any]=512 , UpperCAmelCase_ : Any=2 , UpperCAmelCase_ : Optional[Any]=0.02 , UpperCAmelCase_ : Optional[Any]=1e-1_2 , UpperCAmelCase_ : Any=1 , UpperCAmelCase_ : Dict=0 , UpperCAmelCase_ : Optional[int]=2 , UpperCAmelCase_ : Tuple="absolute" , UpperCAmelCase_ : Union[str, Any]=True , UpperCAmelCase_ : Dict=None , UpperCAmelCase_ : Union[str, Any]=False , UpperCAmelCase_ : int=2 , UpperCAmelCase_ : int=False , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : Union[str, Any]=("en_XX",) , UpperCAmelCase_ : int=None , **UpperCAmelCase_ : str , ): super().__init__(pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = position_embedding_type SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = classifier_dropout SCREAMING_SNAKE_CASE__ = pre_norm SCREAMING_SNAKE_CASE__ = adapter_reduction_factor SCREAMING_SNAKE_CASE__ = adapter_layer_norm SCREAMING_SNAKE_CASE__ = adapter_reuse_layer_norm SCREAMING_SNAKE_CASE__ = ln_before_adapter SCREAMING_SNAKE_CASE__ = list(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = default_language class lowercase__ ( _UpperCAmelCase ): @property def A_ ( self : List[Any] ): if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: SCREAMING_SNAKE_CASE__ = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
176
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase : Dict = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Union[str, Any] = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys _UpperCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
353
'''simple docstring''' from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class a__ ( __A ): """simple docstring""" def __init__(self , __lowercase , __lowercase=None , __lowercase=None , __lowercase=0 ): __lowerCAmelCase = 1.0 if scale is None else scale __lowerCAmelCase = 0.0 if loc is None else loc super().__init__(__lowercase , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=__lowercase )] ) @property def _snake_case (self ): return self.base_dist.mean * self.scale + self.loc @property def _snake_case (self ): return self.base_dist.variance * self.scale**2 @property def _snake_case (self ): return self.variance.sqrt() class a__ ( nn.Module ): """simple docstring""" def __init__(self , __lowercase , __lowercase , __lowercase , **__lowercase ): super().__init__(**__lowercase ) __lowerCAmelCase = args_dim __lowerCAmelCase = nn.ModuleList([nn.Linear(__lowercase , __lowercase ) for dim in args_dim.values()] ) __lowerCAmelCase = domain_map def _snake_case (self , __lowercase ): __lowerCAmelCase = [proj(__lowercase ) for proj in self.proj] return self.domain_map(*__lowercase ) class a__ ( nn.Module ): """simple docstring""" def __init__(self , __lowercase ): super().__init__() __lowerCAmelCase = function def _snake_case (self , __lowercase , *__lowercase ): return self.function(__lowercase , *__lowercase ) class a__ : """simple docstring""" __UpperCamelCase : type __UpperCamelCase : int __UpperCamelCase : Dict[str, int] def __init__(self , __lowercase = 1 ): __lowerCAmelCase = dim __lowerCAmelCase = {k: dim * self.args_dim[k] for k in self.args_dim} def _snake_case (self , __lowercase ): if self.dim == 1: return self.distribution_class(*__lowercase ) else: return Independent(self.distribution_class(*__lowercase ) , 1 ) def _snake_case (self , __lowercase , __lowercase = None , __lowercase = None , ): __lowerCAmelCase = self._base_distribution(__lowercase ) if loc is None and scale is None: return distr else: return AffineTransformed(__lowercase , loc=__lowercase , scale=__lowercase , event_dim=self.event_dim ) @property def _snake_case (self ): return () if self.dim == 1 else (self.dim,) @property def _snake_case (self ): return len(self.event_shape ) @property def _snake_case (self ): return 0.0 def _snake_case (self , __lowercase ): return ParameterProjection( in_features=__lowercase , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def _snake_case (self , *__lowercase ): raise NotImplementedError() @staticmethod def _snake_case (__lowercase ): return (x + torch.sqrt(torch.square(__lowercase ) + 4.0 )) / 2.0 class a__ ( __A ): """simple docstring""" __UpperCamelCase : Dict[str, int] = {"df": 1, "loc": 1, "scale": 1} __UpperCamelCase : type = StudentT @classmethod def _snake_case (cls , __lowercase , __lowercase , __lowercase ): __lowerCAmelCase = cls.squareplus(__lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) __lowerCAmelCase = 2.0 + cls.squareplus(__lowercase ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class a__ ( __A ): """simple docstring""" __UpperCamelCase : Dict[str, int] = {"loc": 1, "scale": 1} __UpperCamelCase : type = Normal @classmethod def _snake_case (cls , __lowercase , __lowercase ): __lowerCAmelCase = cls.squareplus(__lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class a__ ( __A ): """simple docstring""" __UpperCamelCase : Dict[str, int] = {"total_count": 1, "logits": 1} __UpperCamelCase : type = NegativeBinomial @classmethod def _snake_case (cls , __lowercase , __lowercase ): __lowerCAmelCase = cls.squareplus(__lowercase ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def _snake_case (self , __lowercase ): __lowerCAmelCase , __lowerCAmelCase = distr_args if self.dim == 1: return self.distribution_class(total_count=__lowercase , logits=__lowercase ) else: return Independent(self.distribution_class(total_count=__lowercase , logits=__lowercase ) , 1 ) def _snake_case (self , __lowercase , __lowercase = None , __lowercase = None ): __lowerCAmelCase , __lowerCAmelCase = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
9
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __A = logging.get_logger(__name__) class lowerCamelCase__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ): """simple docstring""" warnings.warn( "The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use BeitImageProcessor instead." , _A , ) super().__init__(*_A , **_A )
148
import math import random def A__ ( __lowerCamelCase, __lowerCamelCase = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __UpperCAmelCase = 0.02 def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = float(2 * (random.randint(1, 1_00 )) - 1 ) for _ in range(__lowerCamelCase ): # Forward propagation SCREAMING_SNAKE_CASE_ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? SCREAMING_SNAKE_CASE_ = (expected / 1_00) - layer_a # Error delta SCREAMING_SNAKE_CASE_ = layer_1_error * sigmoid_function(__lowerCamelCase, __lowerCamelCase ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 1_00 if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = int(input("Expected value: ")) __UpperCAmelCase = int(input("Number of propagations: ")) print(forward_propagation(expected, number_propagations))
299
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase : Optional[int] = logging.get_logger(__name__) _lowerCamelCase : Optional[int] = { 'sail/poolformer_s12': 'https://huggingface.co/sail/poolformer_s12/resolve/main/config.json', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __UpperCAmelCase ( _snake_case ): '''simple docstring''' __lowerCAmelCase = '''poolformer''' def __init__(self : List[Any] , _lowerCAmelCase : Optional[Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : Optional[Any]=16 , _lowerCAmelCase : Optional[int]=3 , _lowerCAmelCase : List[Any]=4.0 , _lowerCAmelCase : Optional[int]=[2, 2, 6, 2] , _lowerCAmelCase : str=[64, 128, 320, 512] , _lowerCAmelCase : List[str]=[7, 3, 3, 3] , _lowerCAmelCase : Dict=[4, 2, 2, 2] , _lowerCAmelCase : Union[str, Any]=[2, 1, 1, 1] , _lowerCAmelCase : Tuple=4 , _lowerCAmelCase : Dict=0.0 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=True , _lowerCAmelCase : Optional[int]=1e-5 , _lowerCAmelCase : Union[str, Any]=0.02 , **_lowerCAmelCase : Dict , ): A = num_channels A = patch_size A = stride A = padding A = pool_size A = hidden_sizes A = mlp_ratio A = depths A = patch_sizes A = strides A = num_encoder_blocks A = drop_path_rate A = hidden_act A = use_layer_scale A = layer_scale_init_value A = initializer_range super().__init__(**UpperCamelCase__ ) class __UpperCAmelCase ( _snake_case ): '''simple docstring''' __lowerCAmelCase = version.parse('''1.11''' ) @property def A (self : Dict ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def A (self : List[str] ): return 2e-3
363
'''simple docstring''' import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowerCamelCase : Dict = 'src/diffusers' _lowerCamelCase : Dict = '.' # This is to make sure the diffusers module imported is the one in the repo. _lowerCamelCase : List[str] = importlib.util.spec_from_file_location( 'diffusers', os.path.join(DIFFUSERS_PATH, '__init__.py'), submodule_search_locations=[DIFFUSERS_PATH], ) _lowerCamelCase : Tuple = spec.loader.load_module() def __a ( UpperCAmelCase , UpperCAmelCase ) ->Union[str, Any]: """simple docstring""" return line.startswith(UpperCAmelCase ) or len(UpperCAmelCase ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""" , UpperCAmelCase ) is not None def __a ( UpperCAmelCase ) ->Dict: """simple docstring""" A = object_name.split(""".""" ) A = 0 # First let's find the module where our object lives. A = parts[i] while i < len(UpperCAmelCase ) and not os.path.isfile(os.path.join(UpperCAmelCase , f"""{module}.py""" ) ): i += 1 if i < len(UpperCAmelCase ): A = os.path.join(UpperCAmelCase , parts[i] ) if i >= len(UpperCAmelCase ): raise ValueError(f"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(UpperCAmelCase , f"""{module}.py""" ) , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: A = f.readlines() # Now let's find the class / func in the code! A = """""" A = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase ) and re.search(Rf"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase ): raise ValueError(f""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A = line_index while line_index < len(UpperCAmelCase ) and _should_continue(lines[line_index] , UpperCAmelCase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A = lines[start_index:line_index] return "".join(UpperCAmelCase ) _lowerCamelCase : str = re.compile(R'^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)') _lowerCamelCase : Any = re.compile(R'^\s*(\S+)->(\S+)(\s+.*|$)') _lowerCamelCase : str = re.compile(R'<FILL\s+[^>]*>') def __a ( UpperCAmelCase ) ->str: """simple docstring""" A = code.split("""\n""" ) A = 0 while idx < len(UpperCAmelCase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase ): return re.search(R"""^(\s*)\S""" , lines[idx] ).groups()[0] return "" def __a ( UpperCAmelCase ) ->Optional[int]: """simple docstring""" A = len(get_indent(UpperCAmelCase ) ) > 0 if has_indent: A = f"""class Bla:\n{code}""" A = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=UpperCAmelCase ) A = black.format_str(UpperCAmelCase , mode=UpperCAmelCase ) A , A = style_docstrings_in_code(UpperCAmelCase ) return result[len("""class Bla:\n""" ) :] if has_indent else result def __a ( UpperCAmelCase , UpperCAmelCase=False ) ->List[str]: """simple docstring""" with open(UpperCAmelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: A = f.readlines() A = [] A = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase ): A = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A , A , A = search.groups() A = find_code_in_diffusers(UpperCAmelCase ) A = get_indent(UpperCAmelCase ) A = line_index + 1 if indent == theoretical_indent else line_index + 2 A = theoretical_indent A = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A = True while line_index < len(UpperCAmelCase ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase ): break A = lines[line_index] A = _should_continue(UpperCAmelCase , UpperCAmelCase ) and re.search(f"""^{indent}# End copy""" , UpperCAmelCase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A = lines[start_index:line_index] A = """""".join(UpperCAmelCase ) # Remove any nested `Copied from` comments to avoid circular copies A = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase ) is None] A = """\n""".join(UpperCAmelCase ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase ) > 0: A = replace_pattern.replace("""with""" , """""" ).split(""",""" ) A = [_re_replace_pattern.search(UpperCAmelCase ) for p in patterns] for pattern in patterns: if pattern is None: continue A , A , A = pattern.groups() A = re.sub(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if option.strip() == "all-casing": A = re.sub(obja.lower() , obja.lower() , UpperCAmelCase ) A = re.sub(obja.upper() , obja.upper() , UpperCAmelCase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A = blackify(lines[start_index - 1] + theoretical_code ) A = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A = lines[:start_index] + [theoretical_code] + lines[line_index:] A = start_index + 1 if overwrite and len(UpperCAmelCase ) > 0: # Warn the user a file has been modified. print(f"""Detected changes, rewriting {filename}.""" ) with open(UpperCAmelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(UpperCAmelCase ) return diffs def __a ( UpperCAmelCase = False ) ->int: """simple docstring""" A = glob.glob(os.path.join(UpperCAmelCase , """**/*.py""" ) , recursive=UpperCAmelCase ) A = [] for filename in all_files: A = is_copy_consistent(UpperCAmelCase , UpperCAmelCase ) diffs += [f"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(UpperCAmelCase ) > 0: A = """\n""".join(UpperCAmelCase ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": _lowerCamelCase : List[Any] = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') _lowerCamelCase : Any = parser.parse_args() check_copies(args.fix_and_overwrite)
337
0
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class UpperCamelCase ( lowercase ): UpperCAmelCase : str = CustomTokenizer pass
172
"""simple docstring""" def __UpperCAmelCase ( UpperCAmelCase_ : Tuple ) -> Optional[int]: '''simple docstring''' __snake_case : List[str] = [] __snake_case : Optional[Any] = set({'(', '[', '{'} ) __snake_case : Union[str, Any] = set({')', ']', '}'} ) __snake_case : Tuple = {'{': '}', '[': ']', '(': ')'} for i in range(len(UpperCAmelCase_ ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(UpperCAmelCase_ ) == 0 or (len(UpperCAmelCase_ ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(UpperCAmelCase_ ) == 0 def __UpperCAmelCase ( ) -> Any: '''simple docstring''' __snake_case : Optional[Any] = input('Enter sequence of brackets: ' ) if is_balanced(UpperCAmelCase_ ): print(UpperCAmelCase_ , 'is balanced' ) else: print(UpperCAmelCase_ , 'is not balanced' ) if __name__ == "__main__": main()
172
1
"""simple docstring""" import itertools import math def __UpperCAmelCase ( lowercase ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 ,int(math.sqrt(lowercase ) + 1 ) ,6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __UpperCAmelCase ( ): """simple docstring""" _UpperCAmelCase = 2 while True: if is_prime(lowercase ): yield num num += 1 def __UpperCAmelCase ( lowercase = 1_00_01 ): """simple docstring""" return next(itertools.islice(prime_generator() ,nth - 1 ,lowercase ) ) if __name__ == "__main__": print(F'''{solution() = }''')
30
"""simple docstring""" import string from math import logaa def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" _UpperCAmelCase = document.translate( str.maketrans("""""" ,"""""" ,string.punctuation ) ).replace("""\n""" ,"""""" ) _UpperCAmelCase = document_without_punctuation.split(""" """ ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" _UpperCAmelCase = corpus.lower().translate( str.maketrans("""""" ,"""""" ,string.punctuation ) ) # strip all punctuation and replace it with '' _UpperCAmelCase = corpus_without_punctuation.split("""\n""" ) _UpperCAmelCase = term.lower() return (len([doc for doc in docs if term in doc] ), len(lowercase )) def __UpperCAmelCase ( lowercase ,lowercase ,lowercase=False ): """simple docstring""" if smoothing: if n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(1 + logaa(n / (1 + df) ) ,3 ) if df == 0: raise ZeroDivisionError("""df must be > 0""" ) elif n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(logaa(n / df ) ,3 ) def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return round(tf * idf ,3 )
30
1
snake_case : Dict = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' snake_case : Optional[Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] snake_case : int = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
94
snake_case : str = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' snake_case : List[Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] snake_case : int = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
94
1
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :List[Any] , SCREAMING_SNAKE_CASE :str ) -> Dict: # Construct model if openai_config_file == "": __lowerCAmelCase : List[Any] = OpenAIGPTConfig() else: __lowerCAmelCase : Any = OpenAIGPTConfig.from_json_file(SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = OpenAIGPTModel(SCREAMING_SNAKE_CASE ) # Load weights from numpy load_tf_weights_in_openai_gpt(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model __lowerCAmelCase : Optional[int] = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME __lowerCAmelCase : int = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(SCREAMING_SNAKE_CASE , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--openai_checkpoint_folder_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--openai_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) _UpperCAmelCase = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
232
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = {'vocab_file': 'vocab.txt', 'emoji_file': 'emoji.json'} _UpperCAmelCase = { 'vocab_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt', }, 'emoji_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json', }, } _UpperCAmelCase = { 'abeja/gpt-neox-japanese-2.7b': 2048, } def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :Optional[int] ) -> Optional[Any]: with open(SCREAMING_SNAKE_CASE , """r""" , encoding="""utf-8""" ) as f: __lowerCAmelCase : int = json.loads(f.read() ) __lowerCAmelCase : Dict = collections.OrderedDict() __lowerCAmelCase : str = collections.OrderedDict() __lowerCAmelCase : Union[str, Any] = collections.OrderedDict() with open(SCREAMING_SNAKE_CASE , """r""" , encoding="""utf-8""" ) as f: __lowerCAmelCase : Tuple = f.readlines() __lowerCAmelCase : Tuple = [[t.rstrip("""\n""" )] if (t == """,""" or """,""" not in t) else t.rstrip("""\n""" ).split(""",""" ) for t in token] for idx, b in enumerate(SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Dict = b __lowerCAmelCase : Dict = idx for wd in b: __lowerCAmelCase : List[str] = idx return vocab, raw_vocab, ids_to_tokens, emoji class snake_case_ ( __lowercase ): A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['input_ids', 'attention_mask'] def __init__( self : str , _snake_case : Union[str, Any] , _snake_case : Optional[Any] , _snake_case : Any="<|endoftext|>" , _snake_case : str="<|endoftext|>" , _snake_case : str="<|startoftext|>" , _snake_case : List[Any]="<|endoftext|>" , _snake_case : str=False , **_snake_case : List[Any] , )->Union[str, Any]: '''simple docstring''' super().__init__( unk_token=_snake_case , pad_token=_snake_case , bos_token=_snake_case , eos_token=_snake_case , do_clean_text=_snake_case , **_snake_case , ) if not os.path.isfile(_snake_case ): raise ValueError( F'''Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained''' """ model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) if not os.path.isfile(_snake_case ): raise ValueError( F'''Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google''' """ pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) __lowerCAmelCase : Any = do_clean_text __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = load_vocab_and_emoji(_snake_case , _snake_case ) __lowerCAmelCase : int = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def UpperCAmelCase__ ( self : int )->str: '''simple docstring''' return len(self.raw_vocab ) def UpperCAmelCase__ ( self : Tuple )->Any: '''simple docstring''' return dict(self.raw_vocab , **self.added_tokens_encoder ) def UpperCAmelCase__ ( self : Any , _snake_case : str )->Optional[int]: '''simple docstring''' return self.subword_tokenizer.tokenize(_snake_case , clean=self.do_clean_text ) def UpperCAmelCase__ ( self : Optional[Any] , _snake_case : Optional[Any] )->Any: '''simple docstring''' return self.vocab.get(_snake_case , self.vocab.get(self.unk_token ) ) def UpperCAmelCase__ ( self : int , _snake_case : Any )->int: '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(_snake_case ) def UpperCAmelCase__ ( self : Optional[int] , _snake_case : int )->List[Any]: '''simple docstring''' __lowerCAmelCase : str = """""".join(_snake_case ).strip() return out_string def UpperCAmelCase__ ( self : List[str] , _snake_case : "Conversation" )->List[int]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_snake_case , add_special_tokens=_snake_case ) + [self.eos_token_id] ) if len(_snake_case ) > self.model_max_length: __lowerCAmelCase : List[str] = input_ids[-self.model_max_length :] return input_ids def UpperCAmelCase__ ( self : Optional[Any] , _snake_case : str , _snake_case : Optional[str] = None )->Tuple[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = 0 if os.path.isdir(_snake_case ): __lowerCAmelCase : Dict = os.path.join( _snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCAmelCase : List[Any] = os.path.join( _snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""emoji_file"""] ) else: __lowerCAmelCase : Union[str, Any] = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCAmelCase : Dict = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""emoji_file"""] ) with open(_snake_case , """w""" , encoding="""utf-8""" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' """ Please check that the vocabulary is not corrupted!""" ) __lowerCAmelCase : List[str] = token_index writer.write(""",""".join(_snake_case ) + """\n""" ) index += 1 with open(_snake_case , """w""" , encoding="""utf-8""" ) as writer: json.dump(self.emoji , _snake_case ) return vocab_file, emoji_file class snake_case_ ( __lowercase ): def __init__( self : Optional[Any] , _snake_case : str , _snake_case : Union[str, Any] , _snake_case : Optional[int] )->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = vocab # same as swe __lowerCAmelCase : str = ids_to_tokens # same as bpe __lowerCAmelCase : Dict = emoji __lowerCAmelCase : int = np.max([len(_snake_case ) for w in self.vocab.keys()] ) __lowerCAmelCase : str = re.compile(R"""(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)""" ) __lowerCAmelCase : Optional[Any] = re.compile(R"""[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*""" ) __lowerCAmelCase : Tuple = re.compile(R"""[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}""" ) __lowerCAmelCase : Optional[Any] = re.compile( R"""([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCAmelCase : Union[str, Any] = re.compile( R"""(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCAmelCase : str = re.compile( R"""((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*""" ) __lowerCAmelCase : List[Any] = """─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿""" __lowerCAmelCase : Union[str, Any] = """▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟""" __lowerCAmelCase : str = str.maketrans({k: """<BLOCK>""" for k in keisen + blocks} ) def __len__( self : int )->int: '''simple docstring''' return len(self.ids_to_tokens ) def UpperCAmelCase__ ( self : List[str] , _snake_case : Any )->str: '''simple docstring''' __lowerCAmelCase : List[str] = self.content_repattera.sub("""<URL>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<EMAIL>""" , _snake_case ) __lowerCAmelCase : Optional[Any] = self.content_repattera.sub("""<TEL>""" , _snake_case ) __lowerCAmelCase : str = self.content_repattera.sub("""<DATE>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<DATE>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<PRICE>""" , _snake_case ) __lowerCAmelCase : List[Any] = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __lowerCAmelCase : str = content.replace("""<BLOCK><BLOCK>""" , """<BLOCK>""" ) return content def UpperCAmelCase__ ( self : str , _snake_case : List[Any] , _snake_case : Optional[int]=False )->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = text.replace(""" """ , """<SP>""" ) __lowerCAmelCase : Optional[int] = text.replace(""" """ , """<SP>""" ) __lowerCAmelCase : Union[str, Any] = text.replace("""\r\n""" , """<BR>""" ) __lowerCAmelCase : Tuple = text.replace("""\n""" , """<BR>""" ) __lowerCAmelCase : List[str] = text.replace("""\r""" , """<BR>""" ) __lowerCAmelCase : Dict = text.replace("""\t""" , """<TAB>""" ) __lowerCAmelCase : Dict = text.replace("""—""" , """ー""" ) __lowerCAmelCase : Tuple = text.replace("""−""" , """ー""" ) for k, v in self.emoji["emoji"].items(): if k in text: __lowerCAmelCase : Optional[Any] = text.replace(_snake_case , _snake_case ) if clean: __lowerCAmelCase : List[Any] = self.clean_text(_snake_case ) def check_simbol(_snake_case : List[str] ): __lowerCAmelCase : Optional[int] = x.encode() if len(_snake_case ) == 1 and len(_snake_case ) == 2: __lowerCAmelCase : Optional[Any] = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xc2a1 and c <= 0xc2bf) or (c >= 0xc780 and c <= 0xc783) or (c >= 0xcab9 and c <= 0xcbbf) or (c >= 0xcc80 and c <= 0xcda2) ): return True return False def checkuae(_snake_case : Union[str, Any] ): __lowerCAmelCase : Dict = x.encode() if len(_snake_case ) == 1 and len(_snake_case ) == 3: __lowerCAmelCase : List[str] = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xe2_8080 and c <= 0xe2_b07f: return True return False __lowerCAmelCase : Dict = 0 __lowerCAmelCase : Dict = [] while pos < len(_snake_case ): __lowerCAmelCase : str = min(len(_snake_case ) , pos + self.maxlen + 1 ) if text[pos] == """<""" else pos + 3 __lowerCAmelCase : Tuple = [] # (token_id, token, pos) for e in range(_snake_case , _snake_case , -1 ): __lowerCAmelCase : Optional[int] = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(_snake_case ) > 2: __lowerCAmelCase : Tuple = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(_snake_case ) > 0: # the smallest token_id is adopted __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : int = sorted(_snake_case , key=lambda _snake_case : x[0] )[0] result.append(_snake_case ) __lowerCAmelCase : int = e else: __lowerCAmelCase : Dict = pos + 1 __lowerCAmelCase : Dict = text[pos:end] if check_simbol(_snake_case ): result.append("""<KIGOU>""" ) elif checkuae(_snake_case ): result.append("""<U2000U2BFF>""" ) else: for i in wd.encode("""utf-8""" ): result.append("""<|byte%d|>""" % i ) __lowerCAmelCase : int = end return result def UpperCAmelCase__ ( self : List[str] , _snake_case : Optional[int] , _snake_case : List[Any]="\n" )->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Union[str, Any] = [] __lowerCAmelCase : Optional[Any] = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(_snake_case ) > 0: words.append(bytearray(_snake_case ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCAmelCase : Optional[Any] = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["""emoji_inv"""][word] ) elif word == "<SP>": words.append(""" """ ) elif word == "<BR>": words.append(_snake_case ) elif word == "<TAB>": words.append("""\t""" ) elif word == "<BLOCK>": words.append("""▀""" ) elif word == "<KIGOU>": words.append("""ǀ""" ) elif word == "<U2000U2BFF>": words.append("""‖""" ) else: words.append(_snake_case ) if len(_snake_case ) > 0: words.append(bytearray(_snake_case ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCAmelCase : Dict = """""".join(_snake_case ) return text
232
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase__ = logging.get_logger(__name__) class lowerCamelCase__ ( lowerCAmelCase , lowerCAmelCase): SCREAMING_SNAKE_CASE__ = '''maskformer-swin''' SCREAMING_SNAKE_CASE__ = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__(self , UpperCAmelCase=2_2_4 , UpperCAmelCase=4 , UpperCAmelCase=3 , UpperCAmelCase=9_6 , UpperCAmelCase=[2, 2, 6, 2] , UpperCAmelCase=[3, 6, 1_2, 2_4] , UpperCAmelCase=7 , UpperCAmelCase=4.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase , ) -> Any: super().__init__(**UpperCAmelCase ) _lowercase =image_size _lowercase =patch_size _lowercase =num_channels _lowercase =embed_dim _lowercase =depths _lowercase =len(UpperCAmelCase ) _lowercase =num_heads _lowercase =window_size _lowercase =mlp_ratio _lowercase =qkv_bias _lowercase =hidden_dropout_prob _lowercase =attention_probs_dropout_prob _lowercase =drop_path_rate _lowercase =hidden_act _lowercase =use_absolute_embeddings _lowercase =layer_norm_eps _lowercase =initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowercase =int(embed_dim * 2 ** (len(UpperCAmelCase ) - 1) ) _lowercase =['''stem'''] + [f"stage{idx}" for idx in range(1 , len(UpperCAmelCase ) + 1 )] _lowercase , _lowercase =get_aligned_output_features_output_indices( out_features=UpperCAmelCase , out_indices=UpperCAmelCase , stage_names=self.stage_names )
5
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _SCREAMING_SNAKE_CASE : int = 0 while len(SCREAMING_SNAKE_CASE__ ) > 1: _SCREAMING_SNAKE_CASE : Any = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): _SCREAMING_SNAKE_CASE : Optional[int] = files.index(min(SCREAMING_SNAKE_CASE__ ) ) temp += files[min_index] files.pop(SCREAMING_SNAKE_CASE__ ) files.append(SCREAMING_SNAKE_CASE__ ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
200
0
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = OrderedDict( [ ('''audio-spectrogram-transformer''', '''ASTFeatureExtractor'''), ('''beit''', '''BeitFeatureExtractor'''), ('''chinese_clip''', '''ChineseCLIPFeatureExtractor'''), ('''clap''', '''ClapFeatureExtractor'''), ('''clip''', '''CLIPFeatureExtractor'''), ('''clipseg''', '''ViTFeatureExtractor'''), ('''conditional_detr''', '''ConditionalDetrFeatureExtractor'''), ('''convnext''', '''ConvNextFeatureExtractor'''), ('''cvt''', '''ConvNextFeatureExtractor'''), ('''data2vec-audio''', '''Wav2Vec2FeatureExtractor'''), ('''data2vec-vision''', '''BeitFeatureExtractor'''), ('''deformable_detr''', '''DeformableDetrFeatureExtractor'''), ('''deit''', '''DeiTFeatureExtractor'''), ('''detr''', '''DetrFeatureExtractor'''), ('''dinat''', '''ViTFeatureExtractor'''), ('''donut-swin''', '''DonutFeatureExtractor'''), ('''dpt''', '''DPTFeatureExtractor'''), ('''encodec''', '''EncodecFeatureExtractor'''), ('''flava''', '''FlavaFeatureExtractor'''), ('''glpn''', '''GLPNFeatureExtractor'''), ('''groupvit''', '''CLIPFeatureExtractor'''), ('''hubert''', '''Wav2Vec2FeatureExtractor'''), ('''imagegpt''', '''ImageGPTFeatureExtractor'''), ('''layoutlmv2''', '''LayoutLMv2FeatureExtractor'''), ('''layoutlmv3''', '''LayoutLMv3FeatureExtractor'''), ('''levit''', '''LevitFeatureExtractor'''), ('''maskformer''', '''MaskFormerFeatureExtractor'''), ('''mctct''', '''MCTCTFeatureExtractor'''), ('''mobilenet_v1''', '''MobileNetV1FeatureExtractor'''), ('''mobilenet_v2''', '''MobileNetV2FeatureExtractor'''), ('''mobilevit''', '''MobileViTFeatureExtractor'''), ('''nat''', '''ViTFeatureExtractor'''), ('''owlvit''', '''OwlViTFeatureExtractor'''), ('''perceiver''', '''PerceiverFeatureExtractor'''), ('''poolformer''', '''PoolFormerFeatureExtractor'''), ('''regnet''', '''ConvNextFeatureExtractor'''), ('''resnet''', '''ConvNextFeatureExtractor'''), ('''segformer''', '''SegformerFeatureExtractor'''), ('''sew''', '''Wav2Vec2FeatureExtractor'''), ('''sew-d''', '''Wav2Vec2FeatureExtractor'''), ('''speech_to_text''', '''Speech2TextFeatureExtractor'''), ('''speecht5''', '''SpeechT5FeatureExtractor'''), ('''swiftformer''', '''ViTFeatureExtractor'''), ('''swin''', '''ViTFeatureExtractor'''), ('''swinv2''', '''ViTFeatureExtractor'''), ('''table-transformer''', '''DetrFeatureExtractor'''), ('''timesformer''', '''VideoMAEFeatureExtractor'''), ('''tvlt''', '''TvltFeatureExtractor'''), ('''unispeech''', '''Wav2Vec2FeatureExtractor'''), ('''unispeech-sat''', '''Wav2Vec2FeatureExtractor'''), ('''van''', '''ConvNextFeatureExtractor'''), ('''videomae''', '''VideoMAEFeatureExtractor'''), ('''vilt''', '''ViltFeatureExtractor'''), ('''vit''', '''ViTFeatureExtractor'''), ('''vit_mae''', '''ViTFeatureExtractor'''), ('''vit_msn''', '''ViTFeatureExtractor'''), ('''wav2vec2''', '''Wav2Vec2FeatureExtractor'''), ('''wav2vec2-conformer''', '''Wav2Vec2FeatureExtractor'''), ('''wavlm''', '''Wav2Vec2FeatureExtractor'''), ('''whisper''', '''WhisperFeatureExtractor'''), ('''xclip''', '''CLIPFeatureExtractor'''), ('''yolos''', '''YolosFeatureExtractor'''), ] ) lowerCAmelCase = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def _lowerCamelCase( lowercase__ ) -> Dict: '''simple docstring''' for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: __lowercase= model_type_to_module_name(lowercase__ ) __lowercase= importlib.import_module(F'.{module_name}' , 'transformers.models' ) try: return getattr(lowercase__ , lowercase__ ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(lowercase__ , '__name__' , lowercase__ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __lowercase= importlib.import_module('transformers' ) if hasattr(lowercase__ , lowercase__ ): return getattr(lowercase__ , lowercase__ ) return None def _lowerCamelCase( lowercase__ , lowercase__ = None , lowercase__ = False , lowercase__ = False , lowercase__ = None , lowercase__ = None , lowercase__ = None , lowercase__ = False , **lowercase__ , ) -> List[str]: '''simple docstring''' __lowercase= get_file_from_repo( lowercase__ , lowercase__ , cache_dir=lowercase__ , force_download=lowercase__ , resume_download=lowercase__ , proxies=lowercase__ , use_auth_token=lowercase__ , revision=lowercase__ , local_files_only=lowercase__ , ) if resolved_config_file is None: logger.info( 'Could not locate the feature extractor configuration file, will try to use the model config instead.' ) return {} with open(lowercase__ , encoding='utf-8' ) as reader: return json.load(lowercase__ ) class A : def __init__(self ): raise EnvironmentError( 'AutoFeatureExtractor is designed to be instantiated ' 'using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(lowerCAmelCase ) def _A (cls , lowerCAmelCase , **lowerCAmelCase ): __lowercase= kwargs.pop('config' , lowerCAmelCase ) __lowercase= kwargs.pop('trust_remote_code' , lowerCAmelCase ) __lowercase= True __lowercase, __lowercase= FeatureExtractionMixin.get_feature_extractor_dict(lowerCAmelCase , **lowerCAmelCase ) __lowercase= config_dict.get('feature_extractor_type' , lowerCAmelCase ) __lowercase= None if "AutoFeatureExtractor" in config_dict.get('auto_map' , {} ): __lowercase= config_dict['auto_map']['AutoFeatureExtractor'] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): __lowercase= AutoConfig.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) # It could be in `config.feature_extractor_type`` __lowercase= getattr(lowerCAmelCase , 'feature_extractor_type' , lowerCAmelCase ) if hasattr(lowerCAmelCase , 'auto_map' ) and "AutoFeatureExtractor" in config.auto_map: __lowercase= config.auto_map['AutoFeatureExtractor'] if feature_extractor_class is not None: __lowercase= feature_extractor_class_from_name(lowerCAmelCase ) __lowercase= feature_extractor_auto_map is not None __lowercase= feature_extractor_class is not None or type(lowerCAmelCase ) in FEATURE_EXTRACTOR_MAPPING __lowercase= resolve_trust_remote_code( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if has_remote_code and trust_remote_code: __lowercase= get_class_from_dynamic_module( lowerCAmelCase , lowerCAmelCase , **lowerCAmelCase ) __lowercase= kwargs.pop('code_revision' , lowerCAmelCase ) if os.path.isdir(lowerCAmelCase ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(lowerCAmelCase , **lowerCAmelCase ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(lowerCAmelCase , **lowerCAmelCase ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(lowerCAmelCase ) in FEATURE_EXTRACTOR_MAPPING: __lowercase= FEATURE_EXTRACTOR_MAPPING[type(lowerCAmelCase )] return feature_extractor_class.from_dict(lowerCAmelCase , **lowerCAmelCase ) raise ValueError( f'Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ' f'`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ' f'`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}' ) @staticmethod def _A (lowerCAmelCase , lowerCAmelCase ): FEATURE_EXTRACTOR_MAPPING.register(lowerCAmelCase , lowerCAmelCase )
366
lowerCAmelCase = [ 9_9_9, 8_0_0, 7_9_9, 6_0_0, 5_9_9, 5_0_0, 4_0_0, 3_9_9, 3_7_7, 3_5_5, 3_3_3, 3_1_1, 2_8_8, 2_6_6, 2_4_4, 2_2_2, 2_0_0, 1_9_9, 1_7_7, 1_5_5, 1_3_3, 1_1_1, 8_8, 6_6, 4_4, 2_2, 0, ] lowerCAmelCase = [ 9_9_9, 9_7_6, 9_5_2, 9_2_8, 9_0_5, 8_8_2, 8_5_8, 8_5_7, 8_1_0, 7_6_2, 7_1_5, 7_1_4, 5_7_2, 4_2_9, 4_2_8, 2_8_6, 2_8_5, 2_3_8, 1_9_0, 1_4_3, 1_4_2, 1_1_8, 9_5, 7_1, 4_7, 2_4, 0, ] lowerCAmelCase = [ 9_9_9, 9_8_8, 9_7_7, 9_6_6, 9_5_5, 9_4_4, 9_3_3, 9_2_2, 9_1_1, 9_0_0, 8_9_9, 8_7_9, 8_5_9, 8_4_0, 8_2_0, 8_0_0, 7_9_9, 7_6_6, 7_3_3, 7_0_0, 6_9_9, 6_5_0, 6_0_0, 5_9_9, 5_0_0, 4_9_9, 4_0_0, 3_9_9, 3_5_0, 3_0_0, 2_9_9, 2_6_6, 2_3_3, 2_0_0, 1_9_9, 1_7_9, 1_5_9, 1_4_0, 1_2_0, 1_0_0, 9_9, 8_8, 7_7, 6_6, 5_5, 4_4, 3_3, 2_2, 1_1, 0, ] lowerCAmelCase = [ 9_9_9, 9_9_5, 9_9_2, 9_8_9, 9_8_5, 9_8_1, 9_7_8, 9_7_5, 9_7_1, 9_6_7, 9_6_4, 9_6_1, 9_5_7, 9_5_6, 9_5_1, 9_4_7, 9_4_2, 9_3_7, 9_3_3, 9_2_8, 9_2_3, 9_1_9, 9_1_4, 9_1_3, 9_0_8, 9_0_3, 8_9_7, 8_9_2, 8_8_7, 8_8_1, 8_7_6, 8_7_1, 8_7_0, 8_6_4, 8_5_8, 8_5_2, 8_4_6, 8_4_0, 8_3_4, 8_2_8, 8_2_7, 8_2_0, 8_1_3, 8_0_6, 7_9_9, 7_9_2, 7_8_5, 7_8_4, 7_7_7, 7_7_0, 7_6_3, 7_5_6, 7_4_9, 7_4_2, 7_4_1, 7_3_3, 7_2_4, 7_1_6, 7_0_7, 6_9_9, 6_9_8, 6_8_8, 6_7_7, 6_6_6, 6_5_6, 6_5_5, 6_4_5, 6_3_4, 6_2_3, 6_1_3, 6_1_2, 5_9_8, 5_8_4, 5_7_0, 5_6_9, 5_5_5, 5_4_1, 5_2_7, 5_2_6, 5_0_5, 4_8_4, 4_8_3, 4_6_2, 4_4_0, 4_3_9, 3_9_6, 3_9_5, 3_5_2, 3_5_1, 3_0_8, 3_0_7, 2_6_4, 2_6_3, 2_2_0, 2_1_9, 1_7_6, 1_3_2, 8_8, 4_4, 0, ] lowerCAmelCase = [ 9_9_9, 9_9_7, 9_9_5, 9_9_2, 9_9_0, 9_8_8, 9_8_6, 9_8_4, 9_8_1, 9_7_9, 9_7_7, 9_7_5, 9_7_2, 9_7_0, 9_6_8, 9_6_6, 9_6_4, 9_6_1, 9_5_9, 9_5_7, 9_5_6, 9_5_4, 9_5_1, 9_4_9, 9_4_6, 9_4_4, 9_4_1, 9_3_9, 9_3_6, 9_3_4, 9_3_1, 9_2_9, 9_2_6, 9_2_4, 9_2_1, 9_1_9, 9_1_6, 9_1_4, 9_1_3, 9_1_0, 9_0_7, 9_0_5, 9_0_2, 8_9_9, 8_9_6, 8_9_3, 8_9_1, 8_8_8, 8_8_5, 8_8_2, 8_7_9, 8_7_7, 8_7_4, 8_7_1, 8_7_0, 8_6_7, 8_6_4, 8_6_1, 8_5_8, 8_5_5, 8_5_2, 8_4_9, 8_4_6, 8_4_3, 8_4_0, 8_3_7, 8_3_4, 8_3_1, 8_2_8, 8_2_7, 8_2_4, 8_2_1, 8_1_7, 8_1_4, 8_1_1, 8_0_8, 8_0_4, 8_0_1, 7_9_8, 7_9_5, 7_9_1, 7_8_8, 7_8_5, 7_8_4, 7_8_0, 7_7_7, 7_7_4, 7_7_0, 7_6_6, 7_6_3, 7_6_0, 7_5_6, 7_5_2, 7_4_9, 7_4_6, 7_4_2, 7_4_1, 7_3_7, 7_3_3, 7_3_0, 7_2_6, 7_2_2, 7_1_8, 7_1_4, 7_1_0, 7_0_7, 7_0_3, 6_9_9, 6_9_8, 6_9_4, 6_9_0, 6_8_5, 6_8_1, 6_7_7, 6_7_3, 6_6_9, 6_6_4, 6_6_0, 6_5_6, 6_5_5, 6_5_0, 6_4_6, 6_4_1, 6_3_6, 6_3_2, 6_2_7, 6_2_2, 6_1_8, 6_1_3, 6_1_2, 6_0_7, 6_0_2, 5_9_6, 5_9_1, 5_8_6, 5_8_0, 5_7_5, 5_7_0, 5_6_9, 5_6_3, 5_5_7, 5_5_1, 5_4_5, 5_3_9, 5_3_3, 5_2_7, 5_2_6, 5_1_9, 5_1_2, 5_0_5, 4_9_8, 4_9_1, 4_8_4, 4_8_3, 4_7_4, 4_6_6, 4_5_7, 4_4_9, 4_4_0, 4_3_9, 4_2_8, 4_1_8, 4_0_7, 3_9_6, 3_9_5, 3_8_1, 3_6_6, 3_5_2, 3_5_1, 3_3_0, 3_0_8, 3_0_7, 2_8_6, 2_6_4, 2_6_3, 2_4_2, 2_2_0, 2_1_9, 1_7_6, 1_7_5, 1_3_2, 1_3_1, 8_8, 4_4, 0, ] lowerCAmelCase = [ 9_9_9, 9_9_1, 9_8_2, 9_7_4, 9_6_6, 9_5_8, 9_5_0, 9_4_1, 9_3_3, 9_2_5, 9_1_6, 9_0_8, 9_0_0, 8_9_9, 8_7_4, 8_5_0, 8_2_5, 8_0_0, 7_9_9, 7_0_0, 6_0_0, 5_0_0, 4_0_0, 3_0_0, 2_0_0, 1_0_0, 0, ] lowerCAmelCase = [ 9_9_9, 9_9_2, 9_8_5, 9_7_8, 9_7_1, 9_6_4, 9_5_7, 9_4_9, 9_4_2, 9_3_5, 9_2_8, 9_2_1, 9_1_4, 9_0_7, 9_0_0, 8_9_9, 8_7_9, 8_5_9, 8_4_0, 8_2_0, 8_0_0, 7_9_9, 7_6_6, 7_3_3, 7_0_0, 6_9_9, 6_5_0, 6_0_0, 5_9_9, 5_0_0, 4_9_9, 4_0_0, 3_9_9, 3_0_0, 2_9_9, 2_0_0, 1_9_9, 1_0_0, 9_9, 0, ] lowerCAmelCase = [ 9_9_9, 9_9_6, 9_9_2, 9_8_9, 9_8_5, 9_8_2, 9_7_9, 9_7_5, 9_7_2, 9_6_8, 9_6_5, 9_6_1, 9_5_8, 9_5_5, 9_5_1, 9_4_8, 9_4_4, 9_4_1, 9_3_8, 9_3_4, 9_3_1, 9_2_7, 9_2_4, 9_2_0, 9_1_7, 9_1_4, 9_1_0, 9_0_7, 9_0_3, 9_0_0, 8_9_9, 8_9_1, 8_8_4, 8_7_6, 8_6_9, 8_6_1, 8_5_3, 8_4_6, 8_3_8, 8_3_0, 8_2_3, 8_1_5, 8_0_8, 8_0_0, 7_9_9, 7_8_8, 7_7_7, 7_6_6, 7_5_5, 7_4_4, 7_3_3, 7_2_2, 7_1_1, 7_0_0, 6_9_9, 6_8_8, 6_7_7, 6_6_6, 6_5_5, 6_4_4, 6_3_3, 6_2_2, 6_1_1, 6_0_0, 5_9_9, 5_8_5, 5_7_1, 5_5_7, 5_4_2, 5_2_8, 5_1_4, 5_0_0, 4_9_9, 4_8_5, 4_7_1, 4_5_7, 4_4_2, 4_2_8, 4_1_4, 4_0_0, 3_9_9, 3_7_9, 3_5_9, 3_4_0, 3_2_0, 3_0_0, 2_9_9, 2_7_9, 2_5_9, 2_4_0, 2_2_0, 2_0_0, 1_9_9, 1_6_6, 1_3_3, 1_0_0, 9_9, 6_6, 3_3, 0, ]
304
0
from __future__ import annotations def lowerCamelCase__ ( snake_case_ : float , snake_case_ : float , snake_case_ : float ) -> dict[str, float]: if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0: raise ValueError('''Resistance cannot be negative''' ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
24
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self : Any , a__ : List[Any] , a__ : Dict=13 , a__ : str=32 , a__ : Tuple=3 , a__ : Optional[Any]=4 , a__ : Optional[int]=[10, 20, 30, 40] , a__ : List[Any]=[2, 2, 3, 2] , a__ : List[Any]=True , a__ : int=True , a__ : List[Any]=37 , a__ : Any="gelu" , a__ : int=10 , a__ : Dict=0.0_2 , a__ : Dict=["stage2", "stage3", "stage4"] , a__ : Tuple=[2, 3, 4] , a__ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = image_size __snake_case = num_channels __snake_case = num_stages __snake_case = hidden_sizes __snake_case = depths __snake_case = is_training __snake_case = use_labels __snake_case = intermediate_size __snake_case = hidden_act __snake_case = num_labels __snake_case = initializer_range __snake_case = out_features __snake_case = out_indices __snake_case = scope def a (self : Dict ): """simple docstring""" __snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.num_labels ) __snake_case = self.get_config() return config, pixel_values, labels def a (self : List[str] ): """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=a__ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def a (self : str , a__ : Union[str, Any] , a__ : List[str] , a__ : List[Any] ): """simple docstring""" __snake_case = ConvNextModel(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a (self : Optional[Any] , a__ : List[Any] , a__ : str , a__ : List[Any] ): """simple docstring""" __snake_case = ConvNextForImageClassification(a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a (self : Tuple , a__ : List[Any] , a__ : List[str] , a__ : List[str] ): """simple docstring""" __snake_case = ConvNextBackbone(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __snake_case = None __snake_case = ConvNextBackbone(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def a (self : Tuple ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case = config_and_inputs __snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A_ : Dict = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) A_ : Optional[Any] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) A_ : Dict = True A_ : Optional[Any] = False A_ : int = False A_ : int = False A_ : List[str] = False def a (self : List[str] ): """simple docstring""" __snake_case = ConvNextModelTester(self ) __snake_case = ConfigTester(self , config_class=a__ , has_text_modality=a__ , hidden_size=37 ) def a (self : Tuple ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a (self : str ): """simple docstring""" return @unittest.skip(reason='''ConvNext does not use inputs_embeds''' ) def a (self : int ): """simple docstring""" pass @unittest.skip(reason='''ConvNext does not support input and output embeddings''' ) def a (self : Dict ): """simple docstring""" pass @unittest.skip(reason='''ConvNext does not use feedforward chunking''' ) def a (self : List[Any] ): """simple docstring""" pass def a (self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a__ ) __snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case = [*signature.parameters.keys()] __snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , a__ ) def a (self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def a (self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*a__ ) def a (self : Dict ): """simple docstring""" def check_hidden_states_output(a__ : List[str] , a__ : str , a__ : Tuple ): __snake_case = model_class(a__ ) model.to(a__ ) model.eval() with torch.no_grad(): __snake_case = model(**self._prepare_for_class(a__ , a__ ) ) __snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __snake_case = self.model_tester.num_stages self.assertEqual(len(a__ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = True check_hidden_states_output(a__ , a__ , a__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case = True check_hidden_states_output(a__ , a__ , a__ ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a__ ) @slow def a (self : Any ): """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = ConvNextModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def lowerCamelCase__ ( ) -> List[str]: __snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def a (self : Tuple ): """simple docstring""" return AutoImageProcessor.from_pretrained('''facebook/convnext-tiny-224''' ) if is_vision_available() else None @slow def a (self : Optional[Any] ): """simple docstring""" __snake_case = ConvNextForImageClassification.from_pretrained('''facebook/convnext-tiny-224''' ).to(a__ ) __snake_case = self.default_image_processor __snake_case = prepare_img() __snake_case = image_processor(images=a__ , return_tensors='''pt''' ).to(a__ ) # forward pass with torch.no_grad(): __snake_case = model(**a__ ) # verify the logits __snake_case = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , a__ ) __snake_case = torch.tensor([-0.0_2_6_0, -0.4_7_3_9, 0.1_9_1_1] ).to(a__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a__ , atol=1E-4 ) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase , _UpperCAmelCase ): A_ : Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () A_ : List[Any] = ConvNextConfig A_ : Optional[Any] = False def a (self : Optional[int] ): """simple docstring""" __snake_case = ConvNextModelTester(self )
24
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a__ ( __SCREAMING_SNAKE_CASE ): _a : List[str] = (KDPMaDiscreteScheduler,) _a : Optional[int] = 1_0 def __SCREAMING_SNAKE_CASE( self , **_A ): """simple docstring""" __lowerCAmelCase = { "num_train_timesteps": 1_1_0_0, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**_snake_case ) return config def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=_snake_case ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02] ): self.check_over_configs(beta_start=_snake_case , beta_end=_snake_case ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_snake_case ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_snake_case ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config(prediction_type="v_prediction" ) __lowerCAmelCase = scheduler_class(**_snake_case ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase = sample.to(_snake_case ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase = scheduler.scale_model_input(_snake_case , _snake_case ) __lowerCAmelCase = model(_snake_case , _snake_case ) __lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(_snake_case ) ) __lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_9_3_4E-0_7 ) < 1E-2 assert abs(result_mean.item() - 6.1_1_1_2E-1_0 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.6_9_3_4_2_8_6_5_0_1_7_0_9_7_2E-0_7 ) < 1E-2 assert abs(result_mean.item() - 0.00_02 ) < 1E-3 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if torch_device == "mps": return __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**_snake_case ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase = sample.to(_snake_case ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase = scheduler.scale_model_input(_snake_case , _snake_case ) __lowerCAmelCase = model(_snake_case , _snake_case ) __lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(_snake_case ) ) __lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if torch_device == "mps": return __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**_snake_case ) scheduler.set_timesteps(self.num_inference_steps , device=_snake_case ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter.to(_snake_case ) * scheduler.init_noise_sigma for t in scheduler.timesteps: __lowerCAmelCase = scheduler.scale_model_input(_snake_case , _snake_case ) __lowerCAmelCase = model(_snake_case , _snake_case ) __lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(_snake_case ) ) __lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) if str(_snake_case ).startswith("cpu" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3
351
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class a__ : _a : str = field( default=snake_case__ , metadata={"""help""": """Model type selected in the list: """ + """, """.join(snake_case__ )} ) _a : str = field( default=snake_case__ , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) _a : int = field( default=1_2_8 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _a : int = field( default=1_2_8 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) _a : int = field( default=6_4 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) _a : int = field( default=3_0 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) _a : bool = field( default=snake_case__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) _a : bool = field( default=snake_case__ , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) _a : float = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=2_0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) _a : int = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class a__ ( snake_case__ ): _a : Any = """train""" _a : Union[str, Any] = """dev""" class a__ ( snake_case__ ): _a : SquadDataTrainingArguments _a : List[SquadFeatures] _a : Split _a : bool def __init__( self , _A , _A , _A = None , _A = Split.train , _A = False , _A = None , _A = "pt" , ): """simple docstring""" __lowerCAmelCase = args __lowerCAmelCase = is_language_sensitive __lowerCAmelCase = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_A , _A ): try: __lowerCAmelCase = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) __lowerCAmelCase = mode # Load data features from cache or dataset file __lowerCAmelCase = "v2" if args.version_2_with_negative else "v1" __lowerCAmelCase = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowerCAmelCase = cached_features_file + ".lock" with FileLock(_A ): if os.path.exists(_A ) and not args.overwrite_cache: __lowerCAmelCase = time.time() __lowerCAmelCase = torch.load(_A ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. __lowerCAmelCase = self.old_features["features"] __lowerCAmelCase = self.old_features.get("dataset" , _A ) __lowerCAmelCase = self.old_features.get("examples" , _A ) logger.info( f"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" " future run" ) else: if mode == Split.dev: __lowerCAmelCase = self.processor.get_dev_examples(args.data_dir ) else: __lowerCAmelCase = self.processor.get_train_examples(args.data_dir ) __lowerCAmelCase , __lowerCAmelCase = squad_convert_examples_to_features( examples=self.examples , tokenizer=_A , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_A , ) __lowerCAmelCase = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _A , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ): """simple docstring""" return len(self.features ) def __getitem__( self , _A ): """simple docstring""" __lowerCAmelCase = self.features[i] __lowerCAmelCase = torch.tensor(feature.input_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.attention_mask , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.token_type_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.cls_index , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.p_mask , dtype=torch.float ) __lowerCAmelCase = torch.tensor(feature.is_impossible , dtype=torch.float ) __lowerCAmelCase = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: __lowerCAmelCase = torch.tensor(feature.start_position , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
102
0
from typing import Any class __A: def __init__( self , _snake_case ) -> Optional[Any]: '''simple docstring''' __a = data __a = None def __repr__( self ) -> str: '''simple docstring''' return F"""Node({self.data})""" class __A: def __init__( self ) -> Dict: '''simple docstring''' __a = None def __iter__( self ) -> Any: '''simple docstring''' __a = self.head while node: yield node.data __a = node.next def __len__( self ) -> int: '''simple docstring''' return sum(1 for _ in self ) def __repr__( self ) -> str: '''simple docstring''' return "->".join([str(_snake_case ) for item in self] ) def __getitem__( self , _snake_case ) -> Any: '''simple docstring''' if not 0 <= index < len(self ): raise ValueError('''list index out of range.''' ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self , _snake_case , _snake_case ) -> None: '''simple docstring''' if not 0 <= index < len(self ): raise ValueError('''list index out of range.''' ) __a = self.head for _ in range(_snake_case ): __a = current.next __a = data def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' self.insert_nth(len(self ) , _snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' self.insert_nth(0 , _snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> None: '''simple docstring''' if not 0 <= index <= len(self ): raise IndexError('''list index out of range''' ) __a = Node(_snake_case ) if self.head is None: __a = new_node elif index == 0: __a = self.head # link new_node to head __a = new_node else: __a = self.head for _ in range(index - 1 ): __a = temp.next __a = temp.next __a = new_node def SCREAMING_SNAKE_CASE_ ( self ) -> None: # print every node data '''simple docstring''' print(self ) def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' return self.delete_nth(0 ) def SCREAMING_SNAKE_CASE_ ( self ) -> Any: # delete from tail '''simple docstring''' return self.delete_nth(len(self ) - 1 ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case = 0 ) -> Any: '''simple docstring''' if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError('''List index out of range.''' ) __a = self.head # default first node if index == 0: __a = self.head.next else: __a = self.head for _ in range(index - 1 ): __a = temp.next __a = temp.next __a = temp.next.next return delete_node.data def SCREAMING_SNAKE_CASE_ ( self ) -> bool: '''simple docstring''' return self.head is None def SCREAMING_SNAKE_CASE_ ( self ) -> None: '''simple docstring''' __a = None __a = self.head while current: # Store the current node's next node. __a = current.next # Make the current node's next point backwards __a = prev # Make the previous node be the current node __a = current # Make the current node the next node (to progress iteration) __a = next_node # Return prev in order to put the head at the end __a = prev def __lowerCAmelCase ( ) -> None: __a = LinkedList() assert linked_list.is_empty() is True assert str(a__ ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(a__ ) == i linked_list.insert_nth(a__ , i + 1 ) assert str(a__ ) == "->".join(str(a__ ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(a__ ) == "->".join(str(a__ ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(a__ ) == 9 assert str(a__ ) == "->".join(str(a__ ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): __a = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(a__ ) == "->".join(str(a__ ) for i in range(-8 , 1 ) ) def __lowerCAmelCase ( ) -> None: __a = [ -9, 100, Node(7734_5112 ), '''dlrow olleH''', 7, 5555, 0, -192.55_555, '''Hello, world!''', 77.9, Node(10 ), None, None, 12.20, ] __a = LinkedList() for i in test_input: linked_list.insert_tail(a__ ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(a__ ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head __a = linked_list.delete_head() assert result == -9 assert ( str(a__ ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail __a = linked_list.delete_tail() assert result == 12.2 assert ( str(a__ ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list __a = linked_list.delete_nth(10 ) assert result is None assert ( str(a__ ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node('''Hello again, world!''' ) ) assert ( str(a__ ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(a__ ) assert ( str(a__ ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(a__ ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def __lowerCAmelCase ( ) -> List[Any]: from doctest import testmod testmod() __a = LinkedList() linked_list.insert_head(input('''Inserting 1st at head ''' ).strip() ) linked_list.insert_head(input('''Inserting 2nd at head ''' ).strip() ) print('''\nPrint list:''' ) linked_list.print_list() linked_list.insert_tail(input('''\nInserting 1st at tail ''' ).strip() ) linked_list.insert_tail(input('''Inserting 2nd at tail ''' ).strip() ) print('''\nPrint list:''' ) linked_list.print_list() print('''\nDelete head''' ) linked_list.delete_head() print('''Delete tail''' ) linked_list.delete_tail() print('''\nPrint list:''' ) linked_list.print_list() print('''\nReverse linked list''' ) linked_list.reverse() print('''\nPrint list:''' ) linked_list.print_list() print('''\nString representation of linked list:''' ) print(a__ ) print('''\nReading/changing Node data using indexing:''' ) print(F"""Element at Position 1: {linked_list[1]}""" ) __a = input('''Enter New Value: ''' ).strip() print('''New list:''' ) print(a__ ) print(F"""length of linked_list is : {len(a__ )}""" ) if __name__ == "__main__": main()
6
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path A : Optional[Any] = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(4_2) A : List[str] = {'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} A : Optional[int] = 'zero2' A : str = 'zero3' A : Tuple = [ZEROa, ZEROa] def __lowerCAmelCase ( a__ , a__ , a__ ) -> Tuple: # customize the test name generator function as we want both params to appear in the sub-test # name, as by default it shows only the first param __a = parameterized.to_safe_name('''_'''.join(str(a__ ) for x in param.args ) ) return F"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test A : Union[str, Any] = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class __A( a ): @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Any: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @require_torch_multi_gpu @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> int: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> str: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @require_torch_multi_gpu @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[Any]: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Tuple: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case = 10 , _snake_case = True , _snake_case = True , _snake_case = True , ) -> Any: '''simple docstring''' __a = models[model] __a = self.run_trainer( stage=_snake_case , model_name=_snake_case , eval_steps=_snake_case , num_train_epochs=1 , distributed=_snake_case , fpaa=_snake_case , ) self.do_checks(_snake_case ) return output_dir def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case = 10 , _snake_case = 1 , _snake_case = True , _snake_case = True , ) -> Union[str, Any]: '''simple docstring''' __a = self.get_auto_remove_tmp_dir('''./xxx''' , after=_snake_case ) __a = F""" --model_name_or_path {model_name} --dataset_name hf-internal-testing/librispeech_asr_dummy --dataset_config_name clean --train_split_name validation --validation_split_name validation --output_dir {output_dir} --num_train_epochs {str(_snake_case )} --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --evaluation_strategy steps --learning_rate 5e-4 --warmup_steps 8 --orthography timit --preprocessing_num_workers 1 --group_by_length --freeze_feature_extractor --report_to none --save_steps 0 --eval_steps {eval_steps} --report_to none """.split() if fpaa: args.extend(['''--fp16'''] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __a = F"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() __a = [F"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] __a = self.get_launcher(_snake_case ) __a = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_snake_case , env=self.get_env() ) return output_dir def SCREAMING_SNAKE_CASE_ ( self , _snake_case=False ) -> List[str]: '''simple docstring''' __a = min(2 , get_gpu_count() ) if distributed else 1 return F"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
6
1
"""simple docstring""" from __future__ import annotations def _A ( UpperCamelCase_ : str) -> list[int]: '''simple docstring''' return [ord(UpperCamelCase_) - 96 for elem in plain] def _A ( UpperCamelCase_ : list[int]) -> str: '''simple docstring''' return "".join(chr(elem + 96) for elem in encoded) def _A ( ) -> None: '''simple docstring''' __lowercase = encode(input("-> ").strip().lower()) print("Encoded: ", UpperCamelCase_) print("Decoded:", decode(UpperCamelCase_)) if __name__ == "__main__": main()
362
"""simple docstring""" import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values _a = argparse.ArgumentParser() parser.add_argument('--user', type=str, default='ubuntu') parser.add_argument('--host', type=str, default='localhost') parser.add_argument('--key_path', type=str, default=None) parser.add_argument('--instance', type=str, default='V100:1') parser.add_argument('--provider', type=str, default='cheapest') parser.add_argument('--use_spot', type=bool, default=False) parser.add_argument('--example', type=str, default='pytorch/text-generation/run_generation.py') _a , _a = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('Cannot specify both BYO and on-demand cluster args') _a = rh.cluster( name='rh-cluster', ips=[args.host], ssh_creds={'ssh_user': args.user, 'ssh_private_key': args.key_path} ) else: _a = rh.cluster( name='rh-cluster', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) _a = args.example.rsplit('/', 1)[0] # Set up remote environment cluster.install_packages(['pip:./']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F"pip install -r transformers/examples/{example_dir}/requirements.txt"]) cluster.run(['pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F"python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}"]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
144
0
"""simple docstring""" import string def lowercase ( lowerCAmelCase__ : str ) -> str: __a = '''''' for i in sequence: __a = ord(lowerCAmelCase__ ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def lowercase ( lowerCAmelCase__ : str ) -> str: __a = string.ascii_letters __a = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(lowerCAmelCase__ )] if c in letters else c for c in sequence ) def lowercase ( ) -> None: from timeit import timeit print('''Running performance benchmarks...''' ) __a = '''from string import printable ; from __main__ import atbash, atbash_slow''' print(f'''> atbash_slow(): {timeit('atbash_slow(printable)' , setup=lowerCAmelCase__ )} seconds''' ) print(f'''> atbash(): {timeit('atbash(printable)' , setup=lowerCAmelCase__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'''{example} encrypted in atbash: {atbash(example)}''') benchmark()
45
# HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCAmelCase__ = float('nan') class lowerCAmelCase__ : '''simple docstring''' def __init__( self , __lowerCamelCase) -> Optional[Any]: _A : List[Any] = sys.stdout _A : str = open(__lowerCamelCase , "a") def __getattr__( self , __lowerCamelCase) -> List[str]: return getattr(self.stdout , __lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> str: self.stdout.write(__lowerCamelCase) # strip tqdm codes self.file.write(re.sub(r"^.*\r" , "" , __lowerCamelCase , 0 , re.M)) def _UpperCAmelCase (UpperCamelCase__ : str=80 , UpperCamelCase__ : Tuple=False ): _A : Tuple = [] # deal with critical env vars _A : Dict = ["CUDA_VISIBLE_DEVICES"] for key in env_keys: _A : Optional[int] = os.environ.get(UpperCamelCase__ , UpperCamelCase__ ) if val is not None: cmd.append(f"{key}={val}" ) # python executable (not always needed if the script is executable) _A : Optional[int] = sys.executable if full_python_path else sys.executable.split("/" )[-1] cmd.append(UpperCamelCase__ ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes _A : Tuple = [] _A : Dict = "" while len(UpperCamelCase__ ) > 0: current_line += f"{cmd.pop(0 )} " if len(UpperCamelCase__ ) == 0 or len(UpperCamelCase__ ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(UpperCamelCase__ ) _A : Union[str, Any] = "" return "\\\n".join(UpperCamelCase__ ) def _UpperCAmelCase (UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Tuple ): # unwrap multi-line input _A : Union[str, Any] = re.sub(r"[\\\n]+" , " " , args.base_cmd ) # remove --output_dir if any and set our own _A : int = re.sub("--output_dir\s+[^\s]+" , "" , args.base_cmd ) args.base_cmd += f" --output_dir {output_dir}" # ensure we have --overwrite_output_dir _A : int = re.sub("--overwrite_output_dir\s+" , "" , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _UpperCAmelCase (UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Tuple , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[int] ): # Enable to debug everything but the run itself, to do it fast and see the progress. # This is useful for debugging the output formatting quickly - we can remove it later once # everybody is happy with the output if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 100 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 1_00.2, 55.66_66, 2_22.22_22_22_22] )} , ) _A : Dict = subprocess.run(UpperCamelCase__ , capture_output=UpperCamelCase__ , text=UpperCamelCase__ ) if verbose: print("STDOUT" , result.stdout ) print("STDERR" , result.stderr ) # save the streams _A : Tuple = variation.replace(" " , "-" ) with open(Path(UpperCamelCase__ ) / f"log.{prefix}.stdout.txt" , "w" ) as f: f.write(result.stdout ) with open(Path(UpperCamelCase__ ) / f"log.{prefix}.stderr.txt" , "w" ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print("failed" ) return {target_metric_key: nan} with io.open(f"{output_dir}/all_results.json" , "r" , encoding="utf-8" ) as f: _A : List[str] = json.load(UpperCamelCase__ ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _UpperCAmelCase (UpperCamelCase__ : int , UpperCamelCase__ : Dict , UpperCamelCase__ : Tuple , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : int , UpperCamelCase__ : Tuple , UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Any , ): _A : Union[str, Any] = [] _A : Optional[int] = [] _A : Any = f"{id}: {variation:<{longest_variation_len}}" _A : Dict = f"{preamble}: " _A : Union[str, Any] = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(UpperCamelCase__ ) , desc=UpperCamelCase__ , leave=UpperCamelCase__ ): _A : Optional[Any] = process_run_single( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) _A : Optional[Any] = single_run_metrics[target_metric_key] if not math.isnan(UpperCamelCase__ ): metrics.append(UpperCamelCase__ ) results.append(UpperCamelCase__ ) outcome += "✓" else: outcome += "✘" _A : str = f"\33[2K\r{outcome}" if len(UpperCamelCase__ ) > 0: _A : List[str] = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} _A : Any = round(mean_metrics[target_metric_key] , 2 ) _A : Tuple = f"{outcome} {mean_target}" if len(UpperCamelCase__ ) > 1: results_str += f" {tuple(round(UpperCamelCase__ , 2 ) for x in results )}" print(UpperCamelCase__ ) _A : Optional[int] = variation return mean_metrics else: print(UpperCamelCase__ ) return {variation_key: variation, target_metric_key: nan} def _UpperCAmelCase (): _A : int = torch.cuda.get_device_properties(torch.device("cuda" ) ) return f"\nDatetime : {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S' )}\n\nSoftware:\ntransformers: {transformers.__version__}\ntorch : {torch.__version__}\ncuda : {torch.version.cuda}\npython : {platform.python_version()}\n\nHardware:\n{torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB\n" def _UpperCAmelCase (UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Dict ): _A : Any = pd.DataFrame(UpperCamelCase__ ) _A : List[str] = "variation" _A : List[Any] = "diff_%" _A : int = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan _A : int = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(UpperCamelCase__ ): # as a fallback, use the minimal value as the sentinel _A : List[str] = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(UpperCamelCase__ ): _A : Optional[Any] = df.apply( lambda UpperCamelCase__ : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis="columns" , ) # re-order columns _A : Union[str, Any] = [variation_key, target_metric_key, diff_key, *report_metric_keys] _A : Any = df.reindex(UpperCamelCase__ , axis="columns" ) # reorder cols # capitalize _A : Tuple = df.rename(str.capitalize , axis="columns" ) # make the cols as narrow as possible _A : List[str] = df.rename(lambda UpperCamelCase__ : c.replace("_" , "<br>" ) , axis="columns" ) _A : Union[str, Any] = df.rename(lambda UpperCamelCase__ : c.replace("_" , "\n" ) , axis="columns" ) _A : Optional[int] = ["", "Copy between the cut-here-lines and paste as is to github or a forum"] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=UpperCamelCase__ , floatfmt=".2f" )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=UpperCamelCase__ , floatfmt=".2f" )] print("\n\n".join(UpperCamelCase__ ) ) def _UpperCAmelCase (): _A : int = argparse.ArgumentParser() parser.add_argument( "--base-cmd" , default=UpperCamelCase__ , type=UpperCamelCase__ , required=UpperCamelCase__ , help="Base cmd" , ) parser.add_argument( "--variations" , default=UpperCamelCase__ , type=UpperCamelCase__ , nargs="+" , required=UpperCamelCase__ , help="Multi-dimensional variations, example: '|--fp16|--bf16' '|--tf32'" , ) parser.add_argument( "--base-variation" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="Baseline variation to compare to. if None the minimal target value will be used to compare against" , ) parser.add_argument( "--target-metric-key" , default=UpperCamelCase__ , type=UpperCamelCase__ , required=UpperCamelCase__ , help="Target metric key in output_dir/all_results.json, e.g., train_samples_per_second" , ) parser.add_argument( "--report-metric-keys" , default="" , type=UpperCamelCase__ , help="Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., 'train_loss train_samples" , ) parser.add_argument( "--repeat-times" , default=1 , type=UpperCamelCase__ , help="How many times to re-run each variation - an average will be reported" , ) parser.add_argument( "--output_dir" , default="output_benchmark" , type=UpperCamelCase__ , help="The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked" , ) parser.add_argument( "--verbose" , default=UpperCamelCase__ , action="store_true" , help="Whether to show the outputs of each run or just the benchmark progress" , ) _A : int = parser.parse_args() _A : Union[str, Any] = args.output_dir Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) _A : Tuple = get_base_command(UpperCamelCase__ , UpperCamelCase__ ) # split each dimension into its --foo variations _A : Dict = [list(map(str.strip , re.split(r"\|" , UpperCamelCase__ ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty _A : Union[str, Any] = list(map(str.strip , map(" ".join , itertools.product(*UpperCamelCase__ ) ) ) ) _A : Union[str, Any] = max(len(UpperCamelCase__ ) for x in variations ) # split wanted keys _A : str = args.report_metric_keys.split() # capture prints into a log file for convenience _A : Optional[int] = f"benchmark-report-{datetime.datetime.now().strftime('%Y-%m-%d-%H-%M-%S' )}.txt" print(f"\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt" ) print(f"and this script's output is also piped into {report_fn}" ) _A : Tuple = Tee(UpperCamelCase__ ) print(f"\n*** Running {len(UpperCamelCase__ )} benchmarks:" ) print(f"Base command: {' '.join(UpperCamelCase__ )}" ) _A : str = "variation" _A : Union[str, Any] = [] for id, variation in enumerate(tqdm(UpperCamelCase__ , desc="Total completion: " , leave=UpperCamelCase__ ) ): _A : Dict = base_cmd + variation.split() results.append( process_run( id + 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , args.target_metric_key , UpperCamelCase__ , args.repeat_times , UpperCamelCase__ , args.verbose , ) ) process_results(UpperCamelCase__ , args.target_metric_key , UpperCamelCase__ , args.base_variation , UpperCamelCase__ ) if __name__ == "__main__": main()
11
0
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __snake_case = 16 __snake_case = 32 def __lowerCAmelCase ( lowercase : Accelerator , lowercase : int = 16 , lowercase : str = "bert-base-cased" ) -> Tuple: """simple docstring""" snake_case : Optional[int] = AutoTokenizer.from_pretrained(lowercase ) snake_case : Union[str, Any] = load_dataset("glue" , "mrpc" ) def tokenize_function(lowercase : Dict ): # max_length=None => use the model max length (it's actually the default) snake_case : str = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=lowercase , max_length=lowercase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset snake_case : Any = datasets.map( lowercase , batched=lowercase , remove_columns=["idx", "sentence1", "sentence2"] , load_from_cache_file=lowercase ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : int = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(lowercase : Tuple ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase , padding="max_length" , max_length=128 , return_tensors="pt" ) return tokenizer.pad(lowercase , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. snake_case : Optional[Any] = DataLoader( tokenized_datasets["train"] , shuffle=lowercase , collate_fn=lowercase , batch_size=lowercase ) snake_case : int = DataLoader( tokenized_datasets["validation"] , shuffle=lowercase , collate_fn=lowercase , batch_size=lowercase ) return train_dataloader, eval_dataloader def __lowerCAmelCase ( lowercase : Optional[Any] , lowercase : List[str] ) -> List[Any]: """simple docstring""" snake_case : Dict = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : int = config["lr"] snake_case : List[Any] = int(config["num_epochs"] ) snake_case : Union[str, Any] = int(config["seed"] ) snake_case : Union[str, Any] = int(config["batch_size"] ) snake_case : int = args.model_name_or_path set_seed(lowercase ) snake_case ,snake_case : Any = get_dataloaders(lowercase , lowercase , lowercase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : Any = AutoModelForSequenceClassification.from_pretrained(lowercase , return_dict=lowercase ) # Instantiate optimizer snake_case : Union[str, Any] = ( AdamW if accelerator.state.deepspeed_plugin is None or "optimizer" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) snake_case : Optional[Any] = optimizer_cls(params=model.parameters() , lr=lowercase ) if accelerator.state.deepspeed_plugin is not None: snake_case : Union[str, Any] = accelerator.state.deepspeed_plugin.deepspeed_config[ "gradient_accumulation_steps" ] else: snake_case : Dict = 1 snake_case : Optional[Any] = (len(lowercase ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): snake_case : List[str] = get_linear_schedule_with_warmup( optimizer=lowercase , num_warmup_steps=0 , num_training_steps=lowercase , ) else: snake_case : List[Any] = DummyScheduler(lowercase , total_num_steps=lowercase , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case ,snake_case ,snake_case ,snake_case ,snake_case : Optional[int] = accelerator.prepare( lowercase , lowercase , lowercase , lowercase , lowercase ) # We need to keep track of how many total steps we have iterated over snake_case : Optional[int] = 0 # We also need to keep track of the stating epoch so files are named properly snake_case : Optional[Any] = 0 # Now we train the model snake_case : List[Any] = evaluate.load("glue" , "mrpc" ) snake_case : int = 0 snake_case : Any = {} for epoch in range(lowercase , lowercase ): model.train() for step, batch in enumerate(lowercase ): snake_case : Any = model(**lowercase ) snake_case : List[Any] = outputs.loss snake_case : List[str] = loss / gradient_accumulation_steps accelerator.backward(lowercase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() snake_case : Optional[int] = 0 for step, batch in enumerate(lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : Optional[Any] = model(**lowercase ) snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times snake_case ,snake_case : int = accelerator.gather( (predictions, batch["labels"]) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase ) - 1: snake_case : Any = predictions[: len(eval_dataloader.dataset ) - samples_seen] snake_case : Optional[Any] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase , references=lowercase , ) snake_case : Any = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' , lowercase ) snake_case : Optional[Any] = eval_metric["accuracy"] if best_performance < eval_metric["accuracy"]: snake_case : str = eval_metric["accuracy"] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), F'Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , "all_results.json" ) , "w" ) as f: json.dump(lowercase , lowercase ) def __lowerCAmelCase ( ) -> Tuple: """simple docstring""" snake_case : List[Any] = argparse.ArgumentParser(description="Simple example of training script tracking peak GPU memory usage." ) parser.add_argument( "--model_name_or_path" , type=lowercase , default="bert-base-cased" , help="Path to pretrained model or model identifier from huggingface.co/models." , required=lowercase , ) parser.add_argument( "--output_dir" , type=lowercase , default="." , help="Optional save directory where all checkpoint folders will be stored. Default is the current working directory." , ) parser.add_argument( "--performance_lower_bound" , type=lowercase , default=lowercase , help="Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value." , ) parser.add_argument( "--num_epochs" , type=lowercase , default=3 , help="Number of train epochs." , ) snake_case : Optional[int] = parser.parse_args() snake_case : List[Any] = {"lr": 2e-5, "num_epochs": args.num_epochs, "seed": 42, "batch_size": 16} training_function(lowercase , lowercase ) if __name__ == "__main__": main()
112
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def __lowerCAmelCase ( lowercase : Optional[Any]="ro" , lowercase : Union[str, Any]="en" , lowercase : str="wmt16" , lowercase : Any=None ) -> None: """simple docstring""" try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("run pip install datasets" ) snake_case : Any = F'{src_lang}-{tgt_lang}' print(F'Converting {dataset}-{pair}' ) snake_case : Union[str, Any] = datasets.load_dataset(lowercase , lowercase ) if save_dir is None: snake_case : int = F'{dataset}-{pair}' snake_case : Optional[Any] = Path(lowercase ) save_dir.mkdir(exist_ok=lowercase ) for split in ds.keys(): print(F'Splitting {split} with {ds[split].num_rows} records' ) # to save to val.source, val.target like summary datasets snake_case : Any = "val" if split == "validation" else split snake_case : List[str] = save_dir.joinpath(F'{fn}.source' ) snake_case : int = save_dir.joinpath(F'{fn}.target' ) snake_case : str = src_path.open("w+" ) snake_case : Optional[int] = tgt_path.open("w+" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): snake_case : int = x["translation"] src_fp.write(ex[src_lang] + "\n" ) tgt_fp.write(ex[tgt_lang] + "\n" ) print(F'Saved {dataset} dataset to {save_dir}' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
112
1
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets SCREAMING_SNAKE_CASE :Optional[Any] = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' SCREAMING_SNAKE_CASE :Any = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' SCREAMING_SNAKE_CASE :Optional[Any] = r'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): """simple docstring""" def lowerCAmelCase__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" ), "references": datasets.Value("string" ), } ) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def lowerCAmelCase__ ( self : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any ) -> Optional[int]: """simple docstring""" snake_case_ = 0.0 for i, j in zip(lowerCAmelCase__ , lowerCAmelCase__ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase__ , lowerCAmelCase__ ) else 0.0 snake_case_ = n_correct / len(lowerCAmelCase__ ) return { "accuracy": accuracy, }
159
def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__=False ): if isinstance(lowercase__ , lowercase__ ) and isinstance(lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : List[str] = len(set_a.intersection(lowercase__ ) ) if alternative_union: __SCREAMING_SNAKE_CASE : int = len(lowercase__ ) + len(lowercase__ ) else: __SCREAMING_SNAKE_CASE : int = len(set_a.union(lowercase__ ) ) return intersection / union if isinstance(lowercase__ , (list, tuple) ) and isinstance(lowercase__ , (list, tuple) ): __SCREAMING_SNAKE_CASE : Dict = [element for element in set_a if element in set_b] if alternative_union: __SCREAMING_SNAKE_CASE : Optional[int] = len(lowercase__ ) + len(lowercase__ ) return len(lowercase__ ) / union else: __SCREAMING_SNAKE_CASE : Tuple = set_a + [element for element in set_b if element not in set_a] return len(lowercase__ ) / len(lowercase__ ) return len(lowercase__ ) / len(lowercase__ ) return None if __name__ == "__main__": __lowerCAmelCase : List[Any] ={'a', 'b', 'c', 'd', 'e'} __lowerCAmelCase : Optional[Any] ={'c', 'd', 'e', 'f', 'h', 'i'} print(jaccard_similarity(set_a, set_b))
9
0
"""simple docstring""" import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A ( snake_case :List[Any] ) -> str: # picklable for multiprocessing return x.sum() def A ( snake_case :Optional[Any] ) -> Optional[int]: # picklable for multiprocessing return i + 1 @dataclass class __lowerCAmelCase : lowercase = 42 lowercase = 42 class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 1 __UpperCamelCase = [1, 2] __UpperCamelCase = {'a': 1, 'b': 2} __UpperCamelCase = {'a': [1, 2], 'b': [3, 4]} __UpperCamelCase = {'a': {'1': 1}, 'b': 2} __UpperCamelCase = {'a': 1, 'b': 2, 'c': 3, 'd': 4} __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 2 __UpperCamelCase = [2, 3] __UpperCamelCase = {'a': 2, 'b': 3} __UpperCamelCase = {'a': [2, 3], 'b': [4, 5]} __UpperCamelCase = {'a': {'1': 2}, 'b': 3} __UpperCamelCase = {'a': 2, 'b': 3, 'c': 4, 'd': 5} self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) __UpperCamelCase = 2 self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) __UpperCamelCase = {'a': np.eye(2 ), 'b': np.zeros(3 ), 'c': np.ones(2 )} __UpperCamelCase = {'a': 2, 'b': 0, 'c': 2} __UpperCamelCase = { 'a': np.eye(2 ).astype(__UpperCAmelCase ), 'b': np.zeros(3 ).astype(__UpperCAmelCase ), 'c': np.ones(2 ).astype(__UpperCAmelCase ), } self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , map_numpy=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual( {k: v.tolist() for k, v in map_nested(__UpperCAmelCase , __UpperCAmelCase , map_numpy=__UpperCAmelCase ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(__UpperCAmelCase , __UpperCAmelCase , map_numpy=__UpperCAmelCase , num_proc=__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual( {k: v.tolist() for k, v in map_nested(__UpperCAmelCase , __UpperCAmelCase , map_numpy=__UpperCAmelCase , num_proc=__UpperCAmelCase ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(__UpperCAmelCase ): # can't pickle a local lambda map_nested(lambda __UpperCAmelCase : x + 1 , __UpperCAmelCase , num_proc=__UpperCAmelCase ) def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = {'a': 1, 'b': 2} __UpperCamelCase = {'a': 3, 'b': 4} __UpperCamelCase = {'a': 5, 'b': 6} __UpperCamelCase = sorted([('a', (1, 3, 5)), ('b', (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) ) , __UpperCAmelCase ) def UpperCAmelCase ( self ): '''simple docstring''' class __lowerCAmelCase : lowercase = "bar" __UpperCamelCase = Foo() self.assertEqual(foo.my_attr , 'bar' ) with temporary_assignment(__UpperCAmelCase , 'my_attr' , 'BAR' ): self.assertEqual(foo.my_attr , 'BAR' ) self.assertEqual(foo.my_attr , 'bar' ) @pytest.mark.parametrize( 'iterable_length, num_proc, expected_num_proc' , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (1_6, 1_6, 1_6), (1_6, 1_7, 1_6), (1_7, 1_6, 1_6), ] , ) def A ( snake_case :Any , snake_case :Optional[Any] , snake_case :List[str] ) -> List[str]: with patch('datasets.utils.py_utils._single_map_nested' ) as mock_single_map_nested, patch( 'datasets.parallel.parallel.Pool' ) as mock_multiprocessing_pool: __UpperCamelCase = {f'{i}': i for i in range(snake_case )} __UpperCamelCase = map_nested(lambda snake_case : x + 1_0 , snake_case , num_proc=snake_case , parallel_min_length=1_6 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): @require_tf def UpperCAmelCase ( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers __UpperCamelCase = layers.Dense(2 ) def gen_random_output(): __UpperCamelCase = tf.random.uniform((1, 3) ) return model(__UpperCAmelCase ).numpy() with temp_seed(42 , set_tensorflow=__UpperCAmelCase ): __UpperCamelCase = gen_random_output() with temp_seed(42 , set_tensorflow=__UpperCAmelCase ): __UpperCamelCase = gen_random_output() __UpperCamelCase = gen_random_output() np.testing.assert_equal(__UpperCAmelCase , __UpperCAmelCase ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def UpperCAmelCase ( self ): '''simple docstring''' import torch def gen_random_output(): __UpperCamelCase = torch.nn.Linear(3 , 2 ) __UpperCamelCase = torch.rand(1 , 3 ) return model(__UpperCAmelCase ).detach().numpy() with temp_seed(42 , set_pytorch=__UpperCAmelCase ): __UpperCamelCase = gen_random_output() with temp_seed(42 , set_pytorch=__UpperCAmelCase ): __UpperCamelCase = gen_random_output() __UpperCamelCase = gen_random_output() np.testing.assert_equal(__UpperCAmelCase , __UpperCAmelCase ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def UpperCAmelCase ( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): __UpperCamelCase = gen_random_output() with temp_seed(42 ): __UpperCamelCase = gen_random_output() __UpperCamelCase = gen_random_output() np.testing.assert_equal(__UpperCAmelCase , __UpperCAmelCase ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize('input_data' , [{}] ) def A ( snake_case :Tuple ) -> List[Any]: __UpperCamelCase = NestedDataStructure(snake_case ).data assert output_data == input_data @pytest.mark.parametrize( 'data, expected_output' , [ ({}, []), ([], []), ('foo', ['foo']), (['foo', 'bar'], ['foo', 'bar']), ([['foo', 'bar']], ['foo', 'bar']), ([[['foo'], ['bar']]], ['foo', 'bar']), ([[['foo'], 'bar']], ['foo', 'bar']), ({'a': 1, 'b': 2}, [1, 2]), ({'a': [1, 2], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[[3], [4]]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, [4]]}, [1, 2, 3, 4]), ({'a': {'1': 1}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': [2]}, [1, 2]), ] , ) def A ( snake_case :List[str] , snake_case :Dict ) -> Dict: __UpperCamelCase = NestedDataStructure(snake_case ).flatten() assert output == expected_output def A ( ) -> Optional[Any]: __UpperCamelCase = A(x=1 , y='foobar' ) __UpperCamelCase = {'x': 1, 'y': 'foobar'} assert asdict(snake_case ) == expected_output __UpperCamelCase = {'a': {'b': A(x=1_0 , y='foo' )}, 'c': [A(x=2_0 , y='bar' )]} __UpperCamelCase = {'a': {'b': {'x': 1_0, 'y': 'foo'}}, 'c': [{'x': 2_0, 'y': 'bar'}]} assert asdict(snake_case ) == expected_output with pytest.raises(snake_case ): asdict([1, A(x=1_0 , y='foo' )] ) def A ( snake_case :str ) -> Optional[int]: return text.split() def A ( snake_case :Tuple ) -> Any: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A ( ) -> Optional[Any]: with Pool(2 ) as pool: __UpperCamelCase = list(iflatmap_unordered(snake_case , _split_text , kwargs_iterable=[{'text': 'hello there'}] * 1_0 ) ) assert out.count('hello' ) == 1_0 assert out.count('there' ) == 1_0 assert len(snake_case ) == 2_0 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: __UpperCamelCase = list(iflatmap_unordered(snake_case , _split_text , kwargs_iterable=[{'text': 'hello there'}] * 1_0 ) ) assert out.count('hello' ) == 1_0 assert out.count('there' ) == 1_0 assert len(snake_case ) == 2_0 # check that we get items as fast as possible with Pool(2 ) as pool: __UpperCamelCase = [] for yield_time, content in iflatmap_unordered( snake_case , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{'content': 'a'}, {'content': 'b'}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(snake_case ) assert out.count('a' ) == 2 assert out.count('b' ) == 2 assert len(snake_case ) == 4
263
"""simple docstring""" import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor UpperCamelCase : Optional[int] = logging.get_logger(__name__) class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' warnings.warn( 'The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DonutImageProcessor instead.' , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
263
1
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
101
from __future__ import annotations __a = [] def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->bool: """simple docstring""" for i in range(len(_UpperCamelCase ) ): if board[row][i] == 1: return False for i in range(len(_UpperCamelCase ) ): if board[i][column] == 1: return False for i, j in zip(range(_UpperCamelCase, -1, -1 ), range(_UpperCamelCase, -1, -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(_UpperCamelCase, -1, -1 ), range(_UpperCamelCase, len(_UpperCamelCase ) ) ): if board[i][j] == 1: return False return True def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->bool: """simple docstring""" if row >= len(_UpperCamelCase ): solution.append(_UpperCamelCase ) printboard(_UpperCamelCase ) print() return True for i in range(len(_UpperCamelCase ) ): if is_safe(_UpperCamelCase, _UpperCamelCase, _UpperCamelCase ): lowercase : int = 1 solve(_UpperCamelCase, row + 1 ) lowercase : Tuple = 0 return False def __lowercase ( _UpperCamelCase ) ->None: """simple docstring""" for i in range(len(_UpperCamelCase ) ): for j in range(len(_UpperCamelCase ) ): if board[i][j] == 1: print('''Q''', end=''' ''' ) else: print('''.''', end=''' ''' ) print() # n=int(input("The no. of queens")) __a = 8 __a = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('''The total no. of solutions are :''', len(solution))
337
0
def __lowerCAmelCase ( a__ = 1000 ) -> int: __a , __a = 1, 1 __a = 2 while True: __a = 0 __a = fa + fa __a , __a = fa, f index += 1 for _ in str(a__ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
from string import ascii_uppercase A : Optional[int] = {char: i for i, char in enumerate(ascii_uppercase)} A : Union[str, Any] = dict(enumerate(ascii_uppercase)) def __lowerCAmelCase ( a__ , a__ ) -> str: __a = len(a__ ) __a = 0 while True: if x == i: __a = 0 if len(a__ ) == len(a__ ): break key += key[i] i += 1 return key def __lowerCAmelCase ( a__ , a__ ) -> str: __a = '''''' __a = 0 for letter in message: if letter == " ": cipher_text += " " else: __a = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def __lowerCAmelCase ( a__ , a__ ) -> str: __a = '''''' __a = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: __a = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def __lowerCAmelCase ( ) -> None: __a = '''THE GERMAN ATTACK''' __a = '''SECRET''' __a = generate_key(a__ , a__ ) __a = cipher_text(a__ , a__ ) print(F"""Encrypted Text = {s}""" ) print(F"""Original Text = {original_text(a__ , a__ )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
33
1
import os import pytest from attr import dataclass __a = 'us-east-1' # defaults region @dataclass class lowercase__: """simple docstring""" a :str a :List[str] = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' a :Optional[int] = { 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 500, 'save_steps': 5_500, } a :Any = {**hyperparameters, 'max_steps': 1_000} @property def _lowercase ( self : Any ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def _lowercase ( self : Union[str, Any] ) -> str: return f'''{self.framework}-transfromers-test''' @property def _lowercase ( self : Optional[Any] ) -> str: return f'''./tests/sagemaker/scripts/{self.framework}''' @property def _lowercase ( self : List[str] ) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def a ( snake_case__: Optional[int] ): '''simple docstring''' lowercase_ = SageMakerTestEnvironment(framework=request.cls.framework )
30
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = { 'configuration_rembert': ['REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RemBertConfig', 'RemBertOnnxConfig'] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['RemBertTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['RemBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'RemBertForCausalLM', 'RemBertForMaskedLM', 'RemBertForMultipleChoice', 'RemBertForQuestionAnswering', 'RemBertForSequenceClassification', 'RemBertForTokenClassification', 'RemBertLayer', 'RemBertModel', 'RemBertPreTrainedModel', 'load_tf_weights_in_rembert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRemBertForCausalLM', 'TFRemBertForMaskedLM', 'TFRemBertForMultipleChoice', 'TFRemBertForQuestionAnswering', 'TFRemBertForSequenceClassification', 'TFRemBertForTokenClassification', 'TFRemBertLayer', 'TFRemBertModel', 'TFRemBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
30
1
"""simple docstring""" def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> bool: snake_case_ = len(_SCREAMING_SNAKE_CASE ) snake_case_ = len(_SCREAMING_SNAKE_CASE ) snake_case_ = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] snake_case_ = True for i in range(_SCREAMING_SNAKE_CASE ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: snake_case_ = True if a[i].islower(): snake_case_ = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
233
"""simple docstring""" def _a ( _SCREAMING_SNAKE_CASE ) -> int: assert column_title.isupper() snake_case_ = 0 snake_case_ = len(_SCREAMING_SNAKE_CASE ) - 1 snake_case_ = 0 while index >= 0: snake_case_ = (ord(column_title[index] ) - 64) * pow(26 , _SCREAMING_SNAKE_CASE ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
233
1
import random from .binary_exp_mod import bin_exp_mod def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : List[Any] , _lowerCamelCase : List[str]=1_000) -> Dict: '''simple docstring''' if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd __UpperCamelCase : Union[str, Any] = n - 1 __UpperCamelCase : Tuple = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) __UpperCamelCase : Dict = 0 while count < prec: __UpperCamelCase : List[str] = random.randint(2 , n - 1) __UpperCamelCase : Union[str, Any] = bin_exp_mod(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) if b != 1: __UpperCamelCase : int = True for _ in range(_lowerCamelCase): if b == n - 1: __UpperCamelCase : Tuple = False break __UpperCamelCase : List[Any] = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": lowercase : Union[str, Any] = abs(int(input('Enter bound : ').strip())) print('Here\'s the list of primes:') print(', '.join(str(i) for i in range(n + 1) if is_prime_big(i)))
232
class lowerCamelCase__ : '''simple docstring''' def __init__( self :int ) -> Dict: __UpperCamelCase : Union[str, Any] = {} def _lowerCamelCase ( self :str ) -> None: print(self.vertex ) for i in self.vertex: print(a , " -> " , " -> ".join([str(a ) for j in self.vertex[i]] ) ) def _lowerCamelCase ( self :List[Any] , a :int , a :int ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(a ) else: # else make a new vertex __UpperCamelCase : Optional[Any] = [to_vertex] def _lowerCamelCase ( self :Tuple ) -> None: # visited array for storing already visited nodes __UpperCamelCase : Dict = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(a , a ) def _lowerCamelCase ( self :Any , a :int , a :list ) -> None: # mark start vertex as visited __UpperCamelCase : int = True print(a , end=" " ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(a , a ) if __name__ == "__main__": lowercase : Dict = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('DFS:') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
232
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCamelCase ( _lowercase , _lowercase , unittest.TestCase ): UpperCAmelCase_ = CycleDiffusionPipeline UpperCAmelCase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "negative_prompt", "height", "width", "negative_prompt_embeds", } UpperCAmelCase_ = PipelineTesterMixin.required_optional_params - {"latents"} UpperCAmelCase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"source_prompt"} ) UpperCAmelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case_ (self ) -> List[str]: torch.manual_seed(0 ) UpperCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) UpperCamelCase = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , num_train_timesteps=10_00 , clip_sample=_SCREAMING_SNAKE_CASE , set_alpha_to_one=_SCREAMING_SNAKE_CASE , ) torch.manual_seed(0 ) UpperCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) UpperCamelCase = CLIPTextModel(_SCREAMING_SNAKE_CASE ) UpperCamelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def snake_case_ (self , __a , __a=0 ) -> Tuple: UpperCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(_SCREAMING_SNAKE_CASE ) ).to(_SCREAMING_SNAKE_CASE ) UpperCamelCase = image / 2 + 0.5 if str(_SCREAMING_SNAKE_CASE ).startswith("mps" ): UpperCamelCase = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: UpperCamelCase = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) UpperCamelCase = { '''prompt''': '''An astronaut riding an elephant''', '''source_prompt''': '''An astronaut riding a horse''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''eta''': 0.1, '''strength''': 0.8, '''guidance_scale''': 3, '''source_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def snake_case_ (self ) -> int: UpperCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase = self.get_dummy_components() UpperCamelCase = CycleDiffusionPipeline(**_SCREAMING_SNAKE_CASE ) UpperCamelCase = pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) UpperCamelCase = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE ) UpperCamelCase = pipe(**_SCREAMING_SNAKE_CASE ) UpperCamelCase = output.images UpperCamelCase = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) UpperCamelCase = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def snake_case_ (self ) -> List[Any]: UpperCamelCase = self.get_dummy_components() for name, module in components.items(): if hasattr(_SCREAMING_SNAKE_CASE , "half" ): UpperCamelCase = module.half() UpperCamelCase = CycleDiffusionPipeline(**_SCREAMING_SNAKE_CASE ) UpperCamelCase = pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) UpperCamelCase = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE ) UpperCamelCase = pipe(**_SCREAMING_SNAKE_CASE ) UpperCamelCase = output.images UpperCamelCase = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) UpperCamelCase = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def snake_case_ (self ) -> str: return super().test_save_load_local() @unittest.skip("non-deterministic pipeline" ) def snake_case_ (self ) -> Dict: return super().test_inference_batch_single_identical() @skip_mps def snake_case_ (self ) -> List[str]: return super().test_dict_tuple_outputs_equivalent() @skip_mps def snake_case_ (self ) -> List[Any]: return super().test_save_load_optional_components() @skip_mps def snake_case_ (self ) -> List[Any]: return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class _lowerCamelCase ( unittest.TestCase ): def snake_case_ (self ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case_ (self ) -> Dict: UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/cycle-diffusion/black_colored_car.png" ) UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy" ) UpperCamelCase = init_image.resize((5_12, 5_12) ) UpperCamelCase = '''CompVis/stable-diffusion-v1-4''' UpperCamelCase = DDIMScheduler.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder="scheduler" ) UpperCamelCase = CycleDiffusionPipeline.from_pretrained( _SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , revision="fp16" ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() UpperCamelCase = '''A black colored car''' UpperCamelCase = '''A blue colored car''' UpperCamelCase = torch.manual_seed(0 ) UpperCamelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , source_prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , num_inference_steps=1_00 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) UpperCamelCase = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def snake_case_ (self ) -> Tuple: UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/cycle-diffusion/black_colored_car.png" ) UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy" ) UpperCamelCase = init_image.resize((5_12, 5_12) ) UpperCamelCase = '''CompVis/stable-diffusion-v1-4''' UpperCamelCase = DDIMScheduler.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder="scheduler" ) UpperCamelCase = CycleDiffusionPipeline.from_pretrained(_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() UpperCamelCase = '''A black colored car''' UpperCamelCase = '''A blue colored car''' UpperCamelCase = torch.manual_seed(0 ) UpperCamelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , source_prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , num_inference_steps=1_00 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) UpperCamelCase = output.images assert np.abs(image - expected_image ).max() < 2e-2
371
"""simple docstring""" import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict lowerCAmelCase__ = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = _TestCommandArgs(dataset=_SCREAMING_SNAKE_CASE , all_configs=_SCREAMING_SNAKE_CASE , save_infos=_SCREAMING_SNAKE_CASE ) UpperCamelCase = TestCommand(*_SCREAMING_SNAKE_CASE ) test_command.run() UpperCamelCase = os.path.join(_SCREAMING_SNAKE_CASE , "README.md" ) assert os.path.exists(_SCREAMING_SNAKE_CASE ) UpperCamelCase = DatasetInfosDict.from_directory(_SCREAMING_SNAKE_CASE ) UpperCamelCase = DatasetInfosDict( { "default": DatasetInfo( features=Features( { "tokens": Sequence(Value("string" ) ), "ner_tags": Sequence( ClassLabel(names=["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] ) ), "langs": Sequence(Value("string" ) ), "spans": Sequence(Value("string" ) ), } ) , splits=[ { "name": "train", "num_bytes": 2_351_563, "num_examples": 10_000, }, { "name": "validation", "num_bytes": 238_418, "num_examples": 1_000, }, ] , download_size=3_940_680 , dataset_size=2_589_981 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: UpperCamelCase , UpperCamelCase = getattr(dataset_infos["default"] , _SCREAMING_SNAKE_CASE ), getattr(expected_dataset_infos["default"] , _SCREAMING_SNAKE_CASE ) if key == "num_bytes": assert is_apercent_close(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif key == "splits": assert list(_SCREAMING_SNAKE_CASE ) == list(_SCREAMING_SNAKE_CASE ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
244
0
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a__: Union[str, Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ReformerTokenizer __SCREAMING_SNAKE_CASE = ReformerTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def UpperCamelCase ( self ): super().setUp() A__ = ReformerTokenizer(_A,keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): A__ = '''<s>''' A__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ),_A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ),_A ) def UpperCamelCase ( self ): A__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0],'''<unk>''' ) self.assertEqual(vocab_keys[1],'''<s>''' ) self.assertEqual(vocab_keys[-1],'''j''' ) self.assertEqual(len(_A ),1000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size,1000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return A__ = self.get_tokenizer() A__ = self.get_rust_tokenizer() A__ = '''I was born in 92000, and this is falsé.''' A__ = tokenizer.tokenize(_A ) A__ = rust_tokenizer.tokenize(_A ) self.assertListEqual(_A,_A ) A__ = tokenizer.encode(_A,add_special_tokens=_A ) A__ = rust_tokenizer.encode(_A,add_special_tokens=_A ) self.assertListEqual(_A,_A ) A__ = self.get_rust_tokenizer() A__ = tokenizer.encode(_A ) A__ = rust_tokenizer.encode(_A ) self.assertListEqual(_A,_A ) def UpperCamelCase ( self,__lowerCamelCase=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): A__ = self.rust_tokenizer_class.from_pretrained(_A,**_A ) # Simple input A__ = '''This is a simple input''' A__ = ['''This is a simple input 1''', '''This is a simple input 2'''] A__ = ('''This is a simple input''', '''This is a pair''') A__ = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(_A,tokenizer_r.encode,_A,max_length=_A,padding='''max_length''' ) # Simple input self.assertRaises(_A,tokenizer_r.encode_plus,_A,max_length=_A,padding='''max_length''' ) # Simple input self.assertRaises( _A,tokenizer_r.batch_encode_plus,_A,max_length=_A,padding='''max_length''',) # Pair input self.assertRaises(_A,tokenizer_r.encode,_A,max_length=_A,padding='''max_length''' ) # Pair input self.assertRaises(_A,tokenizer_r.encode_plus,_A,max_length=_A,padding='''max_length''' ) # Pair input self.assertRaises( _A,tokenizer_r.batch_encode_plus,_A,max_length=_A,padding='''max_length''',) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): A__ = ReformerTokenizer(_A,keep_accents=_A ) A__ = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_A,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ),[285, 46, 10, 170, 382],) A__ = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _A,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ],) A__ = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A,[8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4],) A__ = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ],) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def UpperCamelCase ( self ): A__ = '''Hello World!''' A__ = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(_A,self.big_tokenizer.encode(_A ) ) @slow def UpperCamelCase ( self ): A__ = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) A__ = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(_A,self.big_tokenizer.encode(_A ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence A__ = list(self.big_tokenizer.get_vocab().keys() )[:10] A__ = ''' '''.join(_A ) A__ = self.big_tokenizer.encode_plus(_A,return_tensors='''pt''' ) A__ = self.big_tokenizer.batch_encode_plus([sequence, sequence],return_tensors='''pt''' ) A__ = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) A__ = encoded_sequence['''input_ids'''].shape A__ = ReformerModel(_A ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_A ) model(**_A ) @slow def UpperCamelCase ( self ): # fmt: off A__ = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 A__ = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=_A,model_name='''google/reformer-crime-and-punishment''',revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''',padding=_A,sequences=_A,)
193
'''simple docstring''' import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) _UpperCamelCase : Optional[int] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : a_ = field( default=UpperCamelCase , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCamelCase)} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Pretrained config name or path if not the same as model_name"}) a_ = field( default=UpperCamelCase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) a_ = field( default=UpperCamelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class snake_case__ : a_ = field( default=UpperCamelCase , metadata={"help": "The input training data file (a text file)."}) a_ = field( default=UpperCamelCase , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) a_ = field(default=UpperCamelCase , metadata={"help": "Whether ot not to use whole word mask."}) a_ = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) a_ = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) a_ = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) a_ = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"}) def __UpperCAmelCase ( A : DataTrainingArguments , A : PreTrainedTokenizer , A : bool = False , A : Optional[str] = None , ) -> List[Any]: def _dataset(A : Dict , A : str=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('''You need to set world whole masking and mlm to True for Chinese Whole Word Mask''' ) return LineByLineWithRefDataset( tokenizer=A , file_path=A , block_size=args.block_size , ref_path=A , ) return LineByLineTextDataset(tokenizer=A , file_path=A , block_size=args.block_size ) else: return TextDataset( tokenizer=A , file_path=A , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=A , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(A ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def __UpperCAmelCase ( ) -> Optional[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. UpperCAmelCase_ : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : str = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( '''Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ''' '''or remove the --do_eval argument.''' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , A ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: UpperCAmelCase_ : List[str] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCAmelCase_ : List[str] = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: UpperCAmelCase_ : List[Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.tokenizer_name: UpperCAmelCase_ : str = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCAmelCase_ : List[str] = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another''' ''' script, save it,and load it from here, using --tokenizer_name''' ) if model_args.model_name_or_path: UpperCAmelCase_ : str = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=A , cache_dir=model_args.cache_dir , ) else: logger.info('''Training new model from scratch''' ) UpperCAmelCase_ : int = AutoModelWithLMHead.from_config(A ) model.resize_token_embeddings(len(A ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( '''BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the''' '''--mlm flag (masked language modeling).''' ) if data_args.block_size <= 0: UpperCAmelCase_ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: UpperCAmelCase_ : Dict = min(data_args.block_size , tokenizer.max_len ) # Get datasets UpperCAmelCase_ : str = ( get_dataset(A , tokenizer=A , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) UpperCAmelCase_ : Any = ( get_dataset(A , tokenizer=A , evaluate=A , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": UpperCAmelCase_ : Optional[int] = DataCollatorForPermutationLanguageModeling( tokenizer=A , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: UpperCAmelCase_ : Tuple = DataCollatorForWholeWordMask( tokenizer=A , mlm_probability=data_args.mlm_probability ) else: UpperCAmelCase_ : List[str] = DataCollatorForLanguageModeling( tokenizer=A , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer UpperCAmelCase_ : Any = Trainer( model=A , args=A , data_collator=A , train_dataset=A , eval_dataset=A , prediction_loss_only=A , ) # Training if training_args.do_train: UpperCAmelCase_ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=A ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCAmelCase_ : Tuple = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) UpperCAmelCase_ : Dict = trainer.evaluate() UpperCAmelCase_ : Union[str, Any] = math.exp(eval_output['''eval_loss'''] ) UpperCAmelCase_ : Optional[int] = {'''perplexity''': perplexity} UpperCAmelCase_ : int = os.path.join(training_args.output_dir , '''eval_results_lm.txt''' ) if trainer.is_world_master(): with open(A , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' , A , str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) results.update(A ) return results def __UpperCAmelCase ( A : Tuple ) -> Tuple: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
304
0
import unittest from transformers import DonutProcessor lowerCAmelCase = '''naver-clova-ix/donut-base''' class A ( unittest.TestCase ): def _A (self ): __lowercase= DonutProcessor.from_pretrained(lowerCAmelCase ) def _A (self ): __lowercase= { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } __lowercase= ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) __lowercase= self.processor.tokenajson(lowerCAmelCase ) self.assertDictEqual(lowerCAmelCase , lowerCAmelCase )
304
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json''', # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class A ( A_ ): UpperCamelCase_ : Optional[int] ='''blenderbot-small''' UpperCamelCase_ : Optional[Any] =['''past_key_values'''] UpperCamelCase_ : Optional[int] ={'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__(self , lowerCAmelCase=5_0_2_6_5 , lowerCAmelCase=5_1_2 , lowerCAmelCase=8 , lowerCAmelCase=2_0_4_8 , lowerCAmelCase=1_6 , lowerCAmelCase=8 , lowerCAmelCase=2_0_4_8 , lowerCAmelCase=1_6 , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase="gelu" , lowerCAmelCase=5_1_2 , lowerCAmelCase=0.1 , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=0.02 , lowerCAmelCase=1 , lowerCAmelCase=False , lowerCAmelCase=0 , lowerCAmelCase=1 , lowerCAmelCase=2 , lowerCAmelCase=2 , **lowerCAmelCase , ): __lowercase= vocab_size __lowercase= max_position_embeddings __lowercase= d_model __lowercase= encoder_ffn_dim __lowercase= encoder_layers __lowercase= encoder_attention_heads __lowercase= decoder_ffn_dim __lowercase= decoder_layers __lowercase= decoder_attention_heads __lowercase= dropout __lowercase= attention_dropout __lowercase= activation_dropout __lowercase= activation_function __lowercase= init_std __lowercase= encoder_layerdrop __lowercase= decoder_layerdrop __lowercase= use_cache __lowercase= encoder_layers __lowercase= scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , is_encoder_decoder=lowerCAmelCase , decoder_start_token_id=lowerCAmelCase , forced_eos_token_id=lowerCAmelCase , **lowerCAmelCase , ) class A ( A_ ): @property def _A (self ): if self.task in ["default", "seq2seq-lm"]: __lowercase= OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: __lowercase= {0: 'batch'} __lowercase= {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: __lowercase= {0: 'batch', 1: 'decoder_sequence'} __lowercase= {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(lowerCAmelCase , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. __lowercase= OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: __lowercase, __lowercase= self.num_layers for i in range(lowerCAmelCase ): __lowercase= {0: 'batch', 2: 'past_sequence + sequence'} __lowercase= {0: 'batch', 2: 'past_sequence + sequence'} else: __lowercase= OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def _A (self ): if self.task in ["default", "seq2seq-lm"]: __lowercase= super().outputs else: __lowercase= super(lowerCAmelCase , self ).outputs if self.use_past: __lowercase, __lowercase= self.num_layers for i in range(lowerCAmelCase ): __lowercase= {0: 'batch', 2: 'past_sequence + sequence'} __lowercase= {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def _A (self , lowerCAmelCase , lowerCAmelCase = -1 , lowerCAmelCase = -1 , lowerCAmelCase = False , lowerCAmelCase = None , ): __lowercase= self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # Generate decoder inputs __lowercase= seq_length if not self.use_past else 1 __lowercase= self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) __lowercase= {f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} __lowercase= dict(**lowerCAmelCase , **lowerCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __lowercase, __lowercase= common_inputs['input_ids'].shape __lowercase= common_inputs['decoder_input_ids'].shape[1] __lowercase, __lowercase= self.num_attention_heads __lowercase= ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase= decoder_seq_length + 3 __lowercase= ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __lowercase= torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(lowerCAmelCase , lowerCAmelCase )] , dim=1 ) __lowercase= [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __lowercase, __lowercase= self.num_layers __lowercase= min(lowerCAmelCase , lowerCAmelCase ) __lowercase= max(lowerCAmelCase , lowerCAmelCase ) - min_num_layers __lowercase= 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(lowerCAmelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCAmelCase ), torch.zeros(lowerCAmelCase ), torch.zeros(lowerCAmelCase ), torch.zeros(lowerCAmelCase ), ) ) # TODO: test this. __lowercase= encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(lowerCAmelCase , lowerCAmelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCAmelCase ), torch.zeros(lowerCAmelCase )) ) return common_inputs def _A (self , lowerCAmelCase , lowerCAmelCase = -1 , lowerCAmelCase = -1 , lowerCAmelCase = False , lowerCAmelCase = None , ): __lowercase= self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __lowercase, __lowercase= common_inputs['input_ids'].shape # Not using the same length for past_key_values __lowercase= seqlen + 2 __lowercase, __lowercase= self.num_layers __lowercase, __lowercase= self.num_attention_heads __lowercase= ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase= common_inputs['attention_mask'].dtype __lowercase= torch.cat( [common_inputs['attention_mask'], torch.ones(lowerCAmelCase , lowerCAmelCase , dtype=lowerCAmelCase )] , dim=1 ) __lowercase= [ (torch.zeros(lowerCAmelCase ), torch.zeros(lowerCAmelCase )) for _ in range(lowerCAmelCase ) ] return common_inputs def _A (self , lowerCAmelCase , lowerCAmelCase = -1 , lowerCAmelCase = -1 , lowerCAmelCase = False , lowerCAmelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __lowercase= compute_effective_axis_dimension( lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __lowercase= tokenizer.num_special_tokens_to_add(lowerCAmelCase ) __lowercase= compute_effective_axis_dimension( lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCAmelCase ) # Generate dummy inputs according to compute batch and sequence __lowercase= [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size __lowercase= dict(tokenizer(lowerCAmelCase , return_tensors=lowerCAmelCase ) ) return common_inputs def _A (self , lowerCAmelCase , lowerCAmelCase = -1 , lowerCAmelCase = -1 , lowerCAmelCase = False , lowerCAmelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __lowercase= self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCAmelCase , batch_size=lowerCAmelCase , seq_length=lowerCAmelCase , is_pair=lowerCAmelCase , framework=lowerCAmelCase ) elif self.task == "causal-lm": __lowercase= self._generate_dummy_inputs_for_causal_lm( lowerCAmelCase , batch_size=lowerCAmelCase , seq_length=lowerCAmelCase , is_pair=lowerCAmelCase , framework=lowerCAmelCase ) else: __lowercase= self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase , batch_size=lowerCAmelCase , seq_length=lowerCAmelCase , is_pair=lowerCAmelCase , framework=lowerCAmelCase ) return common_inputs def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if self.task in ["default", "seq2seq-lm"]: __lowercase= super()._flatten_past_key_values_(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else: __lowercase= super(lowerCAmelCase , self )._flatten_past_key_values_( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase )
304
1
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class A ( enum.Enum ): __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Union[str, Any] = 1 __UpperCAmelCase : List[str] = 2 @add_end_docstrings(UpperCAmelCase_ ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__(self : Dict , *__UpperCAmelCase : Union[str, Any] , **__UpperCAmelCase : str ) -> List[Any]: """simple docstring""" super().__init__(*__UpperCAmelCase , **__UpperCAmelCase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. UpperCAmelCase__ = None if self.model.config.prefix is not None: UpperCAmelCase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. UpperCAmelCase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = self._sanitize_parameters(prefix=__UpperCAmelCase , **self._forward_params ) UpperCAmelCase__ = {**self._preprocess_params, **preprocess_params} UpperCAmelCase__ = {**self._forward_params, **forward_params} def lowercase_ (self : Any , __UpperCAmelCase : str=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Any=None , __UpperCAmelCase : str=None , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Tuple=None , **__UpperCAmelCase : Optional[int] , ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = {} if prefix is not None: UpperCAmelCase__ = prefix if prefix: UpperCAmelCase__ = self.tokenizer( __UpperCAmelCase , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors=self.framework ) UpperCAmelCase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f"""{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected""" " [None, 'hole']" ) UpperCAmelCase__ = handle_long_generation preprocess_params.update(__UpperCAmelCase ) UpperCAmelCase__ = generate_kwargs UpperCAmelCase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase__ = ReturnType.TENSORS if return_type is not None: UpperCAmelCase__ = return_type if clean_up_tokenization_spaces is not None: UpperCAmelCase__ = clean_up_tokenization_spaces if stop_sequence is not None: UpperCAmelCase__ = self.tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) if len(__UpperCAmelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) UpperCAmelCase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def lowercase_ (self : Union[str, Any] , *__UpperCAmelCase : Tuple , **__UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*__UpperCAmelCase , **__UpperCAmelCase ) def __call__(self : Dict , __UpperCAmelCase : List[Any] , **__UpperCAmelCase : Tuple ) -> List[Any]: """simple docstring""" return super().__call__(__UpperCAmelCase , **__UpperCAmelCase ) def lowercase_ (self : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any]="" , __UpperCAmelCase : Any=None , **__UpperCAmelCase : List[str] ) -> int: """simple docstring""" UpperCAmelCase__ = self.tokenizer( prefix + prompt_text , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors=self.framework ) UpperCAmelCase__ = prompt_text if handle_long_generation == "hole": UpperCAmelCase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: UpperCAmelCase__ = generate_kwargs["max_new_tokens"] else: UpperCAmelCase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: UpperCAmelCase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) UpperCAmelCase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: UpperCAmelCase__ = inputs["attention_mask"][:, -keep_length:] return inputs def lowercase_ (self : List[Any] , __UpperCAmelCase : str , **__UpperCAmelCase : Tuple ) -> Any: """simple docstring""" UpperCAmelCase__ = model_inputs["input_ids"] UpperCAmelCase__ = model_inputs.get("attention_mask" , __UpperCAmelCase ) # Allow empty prompts if input_ids.shape[1] == 0: UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = 1 else: UpperCAmelCase__ = input_ids.shape[0] UpperCAmelCase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. UpperCAmelCase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: UpperCAmelCase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: UpperCAmelCase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length UpperCAmelCase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL UpperCAmelCase__ = self.model.generate(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = generated_sequence.shape[0] if self.framework == "pt": UpperCAmelCase__ = generated_sequence.reshape(__UpperCAmelCase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": UpperCAmelCase__ = tf.reshape(__UpperCAmelCase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[Any]=ReturnType.FULL_TEXT , __UpperCAmelCase : List[Any]=True ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = model_outputs["generated_sequence"][0] UpperCAmelCase__ = model_outputs["input_ids"] UpperCAmelCase__ = model_outputs["prompt_text"] UpperCAmelCase__ = generated_sequence.numpy().tolist() UpperCAmelCase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: UpperCAmelCase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text UpperCAmelCase__ = self.tokenizer.decode( __UpperCAmelCase , skip_special_tokens=__UpperCAmelCase , clean_up_tokenization_spaces=__UpperCAmelCase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=__UpperCAmelCase , clean_up_tokenization_spaces=__UpperCAmelCase , ) ) if return_type == ReturnType.FULL_TEXT: UpperCAmelCase__ = prompt_text + text[prompt_length:] else: UpperCAmelCase__ = text[prompt_length:] UpperCAmelCase__ = {"generated_text": all_text} records.append(__UpperCAmelCase ) return records
65
"""simple docstring""" import numpy as np def lowercase ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) ->Dict: """simple docstring""" __snake_case : Union[str, Any] = int(np.ceil((x_end - xa) / h ) ) __snake_case : Dict = np.zeros((n + 1,) ) __snake_case : List[Any] = ya __snake_case : int = xa for k in range(_snake_case ): __snake_case : Any = f(_snake_case , y[k] ) __snake_case : List[Any] = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) __snake_case : int = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) __snake_case : Optional[int] = f(x + h , y[k] + h * ka ) __snake_case : Optional[int] = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
102
0
from __future__ import annotations def UpperCAmelCase_ ( _A , _A , _A , ): '''simple docstring''' if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif stress < 0: raise ValueError('''Stress cannot be negative''' ) elif tangential_force < 0: raise ValueError('''Tangential Force cannot be negative''' ) elif area < 0: raise ValueError('''Area cannot be negative''' ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
359
from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def UpperCAmelCase_ ( _A , _A , _A ): '''simple docstring''' if not arr: return None, None, 0 if low == high: return low, high, arr[low] SCREAMING_SNAKE_CASE__ = (low + high) // 2 SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = max_subarray(_A , _A , _A ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = max_subarray(_A , mid + 1 , _A ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = max_cross_sum(_A , _A , _A , _A ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def UpperCAmelCase_ ( _A , _A , _A , _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = float('''-inf''' ), -1 SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = float('''-inf''' ), -1 SCREAMING_SNAKE_CASE__ = 0 for i in range(_A , low - 1 , -1 ): summ += arr[i] if summ > left_sum: SCREAMING_SNAKE_CASE__ = summ SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = 0 for i in range(mid + 1 , high + 1 ): summ += arr[i] if summ > right_sum: SCREAMING_SNAKE_CASE__ = summ SCREAMING_SNAKE_CASE__ = i return max_left, max_right, (left_sum + right_sum) def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = [randint(1 , _A ) for _ in range(_A )] SCREAMING_SNAKE_CASE__ = time.time() max_subarray(_A , 0 , input_size - 1 ) SCREAMING_SNAKE_CASE__ = time.time() return end - start def UpperCAmelCase_ ( ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = [10, 1_00, 10_00, 1_00_00, 5_00_00, 10_00_00, 20_00_00, 30_00_00, 40_00_00, 50_00_00] SCREAMING_SNAKE_CASE__ = [time_max_subarray(_A ) for input_size in input_sizes] print('''No of Inputs\t\tTime Taken''' ) for input_size, runtime in zip(_A , _A ): print(_A , '''\t\t''' , _A ) plt.plot(_A , _A ) plt.xlabel('''Number of Inputs''' ) plt.ylabel('''Time taken in seconds''' ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
218
0
from __future__ import annotations from typing import Any class a__ : """simple docstring""" def __init__( self : Any , UpperCAmelCase__ : int ) ->Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = num_of_nodes SCREAMING_SNAKE_CASE : list[list[int]] = [] SCREAMING_SNAKE_CASE : dict[int, int] = {} def _lowercase ( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) ->List[Any]: """simple docstring""" self.m_edges.append([u_node, v_node, weight] ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int ) ->Any: """simple docstring""" if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def _lowercase ( self : Any , UpperCAmelCase__ : int ) ->List[str]: """simple docstring""" if self.m_component[u_node] != u_node: for k in self.m_component: SCREAMING_SNAKE_CASE : Any = self.find_component(snake_case__ ) def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) ->int: """simple docstring""" if component_size[u_node] <= component_size[v_node]: SCREAMING_SNAKE_CASE : Union[str, Any] = v_node component_size[v_node] += component_size[u_node] self.set_component(snake_case__ ) elif component_size[u_node] >= component_size[v_node]: SCREAMING_SNAKE_CASE : Union[str, Any] = self.find_component(snake_case__ ) component_size[u_node] += component_size[v_node] self.set_component(snake_case__ ) def _lowercase ( self : Dict ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = [] SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) SCREAMING_SNAKE_CASE : Tuple = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: SCREAMING_SNAKE_CASE : Union[str, Any] = edge SCREAMING_SNAKE_CASE : int = self.m_component[u] SCREAMING_SNAKE_CASE : Tuple = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): SCREAMING_SNAKE_CASE : str = [u, v, w] for edge in minimum_weight_edge: if isinstance(snake_case__ , snake_case__ ): SCREAMING_SNAKE_CASE : Optional[int] = edge SCREAMING_SNAKE_CASE : str = self.m_component[u] SCREAMING_SNAKE_CASE : List[str] = self.m_component[v] if u_component != v_component: mst_weight += w self.union(snake_case__ , snake_case__ , snake_case__ ) print(f"Added edge [{u} - {v}]\nAdded weight: {w}\n" ) num_of_components -= 1 SCREAMING_SNAKE_CASE : Optional[Any] = [-1] * self.m_num_of_nodes print(f"The total weight of the minimal spanning tree is: {mst_weight}" ) def __lowercase ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
245
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowercase__ ( snake_case__, unittest.TestCase ): _UpperCAmelCase :Union[str, Any] = KandinskyVaaControlnetImgaImgPipeline _UpperCAmelCase :List[Any] = ["image_embeds", "negative_image_embeds", "image", "hint"] _UpperCAmelCase :List[str] = ["image_embeds", "negative_image_embeds", "image", "hint"] _UpperCAmelCase :Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCAmelCase :str = False @property def UpperCAmelCase__ ( self : Tuple ): return 32 @property def UpperCAmelCase__ ( self : List[Any] ): return 32 @property def UpperCAmelCase__ ( self : Dict ): return self.time_input_dim @property def UpperCAmelCase__ ( self : int ): return self.time_input_dim * 4 @property def UpperCAmelCase__ ( self : Optional[int] ): return 100 @property def UpperCAmelCase__ ( self : int ): torch.manual_seed(0 ) lowerCamelCase_ : Union[str, Any] ={ "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } lowerCamelCase_ : Union[str, Any] =UNetaDConditionModel(**snake_case__ ) return model @property def UpperCAmelCase__ ( self : Any ): return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def UpperCAmelCase__ ( self : int ): torch.manual_seed(0 ) lowerCamelCase_ : int =VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : Optional[int] =self.dummy_unet lowerCamelCase_ : Optional[Any] =self.dummy_movq lowerCamelCase_ : Optional[Any] ={ "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.00_085, "beta_end": 0.012, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } lowerCamelCase_ : Optional[Any] =DDIMScheduler(**snake_case__ ) lowerCamelCase_ : Optional[Any] ={ "unet": unet, "scheduler": scheduler, "movq": movq, } return components def UpperCAmelCase__ ( self : Union[str, Any] , snake_case__ : str , snake_case__ : str=0 ): lowerCamelCase_ : Optional[int] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) lowerCamelCase_ : Optional[Any] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( snake_case__ ) # create init_image lowerCamelCase_ : List[Any] =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) lowerCamelCase_ : Union[str, Any] =image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase_ : Tuple =Image.fromarray(np.uinta(snake_case__ ) ).convert("RGB" ).resize((256, 256) ) # create hint lowerCamelCase_ : Dict =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("mps" ): lowerCamelCase_ : List[Any] =torch.manual_seed(snake_case__ ) else: lowerCamelCase_ : List[str] =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) lowerCamelCase_ : Dict ={ "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : Any ="cpu" lowerCamelCase_ : Dict =self.get_dummy_components() lowerCamelCase_ : Dict =self.pipeline_class(**snake_case__ ) lowerCamelCase_ : str =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Optional[Any] =pipe(**self.get_dummy_inputs(snake_case__ ) ) lowerCamelCase_ : Dict =output.images lowerCamelCase_ : Dict =pipe( **self.get_dummy_inputs(snake_case__ ) , return_dict=snake_case__ , )[0] lowerCamelCase_ : List[str] =image[0, -3:, -3:, -1] lowerCamelCase_ : Optional[int] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ : Union[str, Any] =np.array( [0.54_985_034, 0.55_509_365, 0.52_561_504, 0.5_570_494, 0.5_593_818, 0.5_263_979, 0.50_285_643, 0.5_069_846, 0.51_196_736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def UpperCAmelCase__ ( self : str ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : int ): lowerCamelCase_ : List[Any] =load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy" ) lowerCamelCase_ : Optional[int] =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) lowerCamelCase_ : Optional[int] =init_image.resize((512, 512) ) lowerCamelCase_ : int =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png" ) lowerCamelCase_ : Any =torch.from_numpy(np.array(snake_case__ ) ).float() / 255.0 lowerCamelCase_ : Union[str, Any] =hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowerCamelCase_ : str ="A robot, 4k photo" lowerCamelCase_ : List[Any] =KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(snake_case__ ) lowerCamelCase_ : Any =KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa ) lowerCamelCase_ : List[str] =pipeline.to(snake_case__ ) pipeline.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Tuple =torch.Generator(device="cpu" ).manual_seed(0 ) lowerCamelCase_ , lowerCamelCase_ : Tuple =pipe_prior( snake_case__ , image=snake_case__ , strength=0.85 , generator=snake_case__ , negative_prompt="" , ).to_tuple() lowerCamelCase_ : str =pipeline( image=snake_case__ , image_embeds=snake_case__ , negative_image_embeds=snake_case__ , hint=snake_case__ , generator=snake_case__ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) lowerCamelCase_ : Optional[Any] =output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ )
144
0
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer _lowerCAmelCase : List[Any] = logging.get_logger(__name__) _lowerCAmelCase : Optional[int] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} _lowerCAmelCase : List[str] = { "vocab_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json" }, "merges_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt" }, } _lowerCAmelCase : str = {"allegro/herbert-base-cased": 514} _lowerCAmelCase : Optional[Any] = {} class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = HerbertTokenizer def __init__( self , __snake_case=None , __snake_case=None , __snake_case=None , __snake_case="<s>" , __snake_case="<unk>" , __snake_case="<pad>" , __snake_case="<mask>" , __snake_case="</s>" , **__snake_case , ) -> Optional[Any]: '''simple docstring''' super().__init__( __snake_case , __snake_case , tokenizer_file=__snake_case , cls_token=__snake_case , unk_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sep_token=__snake_case , **__snake_case , ) def __magic_name__ ( self , __snake_case , __snake_case = None ) -> List[int]: '''simple docstring''' __a =[self.cls_token_id] __a =[self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __magic_name__ ( self , __snake_case , __snake_case = None , __snake_case = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1] + ([0] * len(__snake_case )) + [1] def __magic_name__ ( self , __snake_case , __snake_case = None ) -> List[int]: '''simple docstring''' __a =[self.sep_token_id] __a =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __magic_name__ ( self , __snake_case , __snake_case = None ) -> Tuple[str]: '''simple docstring''' __a =self._tokenizer.model.save(__snake_case , name=__snake_case ) return tuple(__snake_case )
350
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase : Tuple = { "configuration_megatron_bert": ["MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegatronBertConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : List[str] = [ "MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MegatronBertForCausalLM", "MegatronBertForMaskedLM", "MegatronBertForMultipleChoice", "MegatronBertForNextSentencePrediction", "MegatronBertForPreTraining", "MegatronBertForQuestionAnswering", "MegatronBertForSequenceClassification", "MegatronBertForTokenClassification", "MegatronBertModel", "MegatronBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys _lowerCAmelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
308
0
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class _UpperCamelCase ( unittest.TestCase , lowerCamelCase__ ): '''simple docstring''' def UpperCamelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = load_tool("""text-to-speech""" ) self.tool.setup() def UpperCamelCase__ ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : int = self.tool("""hey""" ) __SCREAMING_SNAKE_CASE : Union[str, Any] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) ) def UpperCamelCase__ ( self : int ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.tool("""hey""" ) __SCREAMING_SNAKE_CASE : List[Any] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) )
112
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu UpperCamelCase__ : List[Any] = [ '''EAGER''', '''AOT_EAGER''', '''INDUCTOR''', '''NVFUSER''', '''AOT_NVFUSER''', '''AOT_CUDAGRAPHS''', '''OFI''', '''FX2TRT''', '''ONNXRT''', '''IPEX''', ] def lowerCAmelCase_ ( _lowerCamelCase: str , _lowerCamelCase: Union[str, Any]=None , _lowerCamelCase: Optional[int]=None , _lowerCamelCase: str=None ): __SCREAMING_SNAKE_CASE : Optional[int] = True while ask_again: __SCREAMING_SNAKE_CASE : Tuple = input(_lowerCamelCase ) try: if default is not None and len(_lowerCamelCase ) == 0: return default return convert_value(_lowerCamelCase ) if convert_value is not None else result except Exception: if error_message is not None: print(_lowerCamelCase ) def lowerCAmelCase_ ( _lowerCamelCase: Tuple , _lowerCamelCase: Union[str, Any]=[] , _lowerCamelCase: List[Any]=None , _lowerCamelCase: Optional[Any]=0 ): __SCREAMING_SNAKE_CASE : Union[str, Any] = BulletMenu(_lowerCamelCase , _lowerCamelCase ) __SCREAMING_SNAKE_CASE : Dict = menu.run(default_choice=_lowerCamelCase ) return convert_value(_lowerCamelCase ) if convert_value is not None else result def lowerCAmelCase_ ( _lowerCamelCase: Optional[Any] ): __SCREAMING_SNAKE_CASE : List[str] = int(_lowerCamelCase ) return ComputeEnvironment(["""LOCAL_MACHINE""", """AMAZON_SAGEMAKER"""][value] ) def lowerCAmelCase_ ( _lowerCamelCase: Any ): __SCREAMING_SNAKE_CASE : str = int(_lowerCamelCase ) return DistributedType(["""NO""", """MULTI_CPU""", """MULTI_XPU""", """MULTI_GPU""", """MULTI_NPU""", """TPU"""][value] ) def lowerCAmelCase_ ( _lowerCamelCase: Tuple ): __SCREAMING_SNAKE_CASE : Tuple = int(_lowerCamelCase ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def lowerCAmelCase_ ( _lowerCamelCase: Union[str, Any] ): __SCREAMING_SNAKE_CASE : List[str] = int(_lowerCamelCase ) return PrecisionType(["""no""", """fp16""", """bf16""", """fp8"""][value] ) def lowerCAmelCase_ ( _lowerCamelCase: Tuple ): __SCREAMING_SNAKE_CASE : int = int(_lowerCamelCase ) return SageMakerDistributedType(["""NO""", """DATA_PARALLEL""", """MODEL_PARALLEL"""][value] ) def lowerCAmelCase_ ( _lowerCamelCase: List[Any] ): return {"yes": True, "no": False}[value.lower()] class _UpperCamelCase ( argparse.RawDescriptionHelpFormatter ): '''simple docstring''' def UpperCamelCase__ ( self : Tuple , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = super()._format_usage(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = usage.replace("""<command> [<args>] """ , """""" ) return usage
112
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowercase : Union[str, Any] = None lowercase : Union[str, Any] = logging.get_logger(__name__) lowercase : Dict = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowercase : Union[str, Any] = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowercase : Dict = { """google/rembert""": 2_5_6, } lowercase : List[str] = """▁""" class A__ ( _a ): """simple docstring""" __A : str = VOCAB_FILES_NAMES __A : Dict = PRETRAINED_VOCAB_FILES_MAP __A : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : Dict = RemBertTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase=True , lowercase=False , lowercase="[CLS]" , lowercase="[SEP]" , lowercase="<unk>" , lowercase="[SEP]" , lowercase="<pad>" , lowercase="[CLS]" , lowercase="[MASK]" , **lowercase , ) -> str: '''simple docstring''' a__ : str = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_) if isinstance(snake_case_ , snake_case_) else mask_token super().__init__( snake_case_ , tokenizer_file=snake_case_ , do_lower_case=snake_case_ , remove_space=snake_case_ , keep_accents=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , pad_token=snake_case_ , cls_token=snake_case_ , mask_token=snake_case_ , **snake_case_ , ) a__ : List[Any] = do_lower_case a__ : List[str] = remove_space a__ : int = keep_accents a__ : Optional[int] = vocab_file a__ : List[str] = False if not self.vocab_file else True def __lowercase ( self , lowercase , lowercase = None) -> Optional[int]: '''simple docstring''' a__ : Optional[int] = [self.sep_token_id] a__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __lowercase ( self , lowercase , lowercase = None , lowercase = False) -> Optional[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.') return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(snake_case_)) + [1] + ([0] * len(snake_case_)) + [1] return [1] + ([0] * len(snake_case_)) + [1] def __lowercase ( self , lowercase , lowercase = None) -> Dict: '''simple docstring''' a__ : Optional[Any] = [self.sep_token_id] a__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def __lowercase ( self , lowercase , lowercase = None) -> List[Any]: '''simple docstring''' if not os.path.isdir(snake_case_): logger.error('Vocabulary path ({}) should be a directory'.format(snake_case_)) return a__ : Tuple = os.path.join( snake_case_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(snake_case_): copyfile(self.vocab_file , snake_case_) return (out_vocab_file,)
371
def A_ ( A__ ) -> List[str]: # noqa: E741 a__ : Dict = len(A__ ) a__ : str = 0 a__ : Any = [0] * n a__ : int = [False] * n a__ : Optional[Any] = [False] * n def dfs(A__ , A__ , A__ , A__ ): if parent == root: out_edge_count += 1 a__ : Union[str, Any] = True a__ : Optional[Any] = at for to in l[at]: if to == parent: pass elif not visited[to]: a__ : List[Any] = dfs(A__ , A__ , A__ , A__ ) a__ : Dict = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: a__ : Dict = True # AP found via cycle if at == low[to]: a__ : List[Any] = True else: a__ : Optional[int] = min(low[at] , A__ ) return out_edge_count for i in range(A__ ): if not visited[i]: a__ : Tuple = 0 a__ : Any = dfs(A__ , A__ , -1 , A__ ) a__ : List[Any] = out_edge_count > 1 for x in range(len(A__ ) ): if is_art[x] is True: print(A__ ) # Adjacency list of graph lowercase : List[Any] = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
225
0
"""simple docstring""" import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _UpperCAmelCase ( a ,unittest.TestCase ): '''simple docstring''' a__ =GPTaTokenizer a__ =GPTaTokenizerFast a__ =True a__ ={'''add_prefix_space''': True} a__ =False def __lowerCAmelCase ( self ) -> Optional[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCAmelCase : Union[str, Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] _UpperCAmelCase : List[str] = dict(zip(A , range(len(A ) ) ) ) _UpperCAmelCase : int = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] _UpperCAmelCase : Dict = {'''unk_token''': '''<unk>'''} _UpperCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCAmelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A ) ) def __lowerCAmelCase ( self , **A ) -> Dict: kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **A ) def __lowerCAmelCase ( self , **A ) -> List[str]: kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **A ) def __lowerCAmelCase ( self , A ) -> List[Any]: _UpperCAmelCase : List[Any] = '''lower newer''' _UpperCAmelCase : List[str] = '''lower newer''' return input_text, output_text def __lowerCAmelCase ( self ) -> str: _UpperCAmelCase : Union[str, Any] = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _UpperCAmelCase : Optional[Any] = '''lower newer''' _UpperCAmelCase : str = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] _UpperCAmelCase : List[str] = tokenizer.tokenize(A , add_prefix_space=A ) self.assertListEqual(A , A ) _UpperCAmelCase : Tuple = tokens + [tokenizer.unk_token] _UpperCAmelCase : Optional[int] = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , A ) def __lowerCAmelCase ( self ) -> str: if not self.test_rust_tokenizer: return _UpperCAmelCase : Union[str, Any] = self.get_tokenizer() _UpperCAmelCase : Union[str, Any] = self.get_rust_tokenizer(add_prefix_space=A ) _UpperCAmelCase : Any = '''lower newer''' # Testing tokenization _UpperCAmelCase : int = tokenizer.tokenize(A , add_prefix_space=A ) _UpperCAmelCase : List[str] = rust_tokenizer.tokenize(A ) self.assertListEqual(A , A ) # Testing conversion to ids without special tokens _UpperCAmelCase : Optional[int] = tokenizer.encode(A , add_special_tokens=A , add_prefix_space=A ) _UpperCAmelCase : List[str] = rust_tokenizer.encode(A , add_special_tokens=A ) self.assertListEqual(A , A ) # Testing conversion to ids with special tokens _UpperCAmelCase : Optional[Any] = self.get_rust_tokenizer(add_prefix_space=A ) _UpperCAmelCase : List[str] = tokenizer.encode(A , add_prefix_space=A ) _UpperCAmelCase : Optional[int] = rust_tokenizer.encode(A ) self.assertListEqual(A , A ) # Testing the unknown token _UpperCAmelCase : List[str] = tokens + [rust_tokenizer.unk_token] _UpperCAmelCase : List[str] = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A ) , A ) def __lowerCAmelCase ( self , *A , **A ) -> Tuple: # It's very difficult to mix/test pretokenization with byte-level # And get both GPT2 and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def __lowerCAmelCase ( self , A=1_5 ) -> List[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): _UpperCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(A , **A ) # Simple input _UpperCAmelCase : Optional[int] = '''This is a simple input''' _UpperCAmelCase : Optional[Any] = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCAmelCase : Dict = ('''This is a simple input''', '''This is a pair''') _UpperCAmelCase : Optional[int] = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding='''max_length''' ) # Simple input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding='''max_length''' ) # Simple input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding='''max_length''' , ) # Pair input self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding='''max_length''' ) # Pair input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding='''max_length''' ) # Pair input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding='''max_length''' , ) def __lowerCAmelCase ( self ) -> Optional[Any]: _UpperCAmelCase : str = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input _UpperCAmelCase : Optional[int] = '''This is a simple input''' _UpperCAmelCase : List[str] = ['''This is a simple input looooooooong''', '''This is a simple input'''] _UpperCAmelCase : List[Any] = ('''This is a simple input''', '''This is a pair''') _UpperCAmelCase : List[str] = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] _UpperCAmelCase : int = tokenizer.pad_token_id _UpperCAmelCase : str = tokenizer(A , padding='''max_length''' , max_length=3_0 , return_tensors='''np''' ) _UpperCAmelCase : List[str] = tokenizer(A , padding=A , truncate=A , return_tensors='''np''' ) _UpperCAmelCase : List[str] = tokenizer(*A , padding='''max_length''' , max_length=6_0 , return_tensors='''np''' ) _UpperCAmelCase : Any = tokenizer(A , padding=A , truncate=A , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def __lowerCAmelCase ( self ) -> Any: _UpperCAmelCase : Tuple = '''$$$''' _UpperCAmelCase : Tuple = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=A , add_bos_token=A ) _UpperCAmelCase : Optional[Any] = '''This is a simple input''' _UpperCAmelCase : Tuple = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCAmelCase : str = tokenizer.bos_token_id _UpperCAmelCase : List[Any] = tokenizer(A ) _UpperCAmelCase : Union[str, Any] = tokenizer(A ) self.assertEqual(out_s.input_ids[0] , A ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _UpperCAmelCase : Dict = tokenizer.decode(out_s.input_ids ) _UpperCAmelCase : Tuple = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def __lowerCAmelCase ( self ) -> Tuple: pass def __lowerCAmelCase ( self ) -> List[Any]: # TODO: change to self.get_tokenizers() when the fast version is implemented _UpperCAmelCase : Any = [self.get_tokenizer(do_lower_case=A , add_bos_token=A )] for tokenizer in tokenizers: with self.subTest(f'{tokenizer.__class__.__name__}' ): _UpperCAmelCase : List[str] = '''Encode this.''' _UpperCAmelCase : List[Any] = '''This one too please.''' _UpperCAmelCase : str = tokenizer.encode(A , add_special_tokens=A ) encoded_sequence += tokenizer.encode(A , add_special_tokens=A ) _UpperCAmelCase : List[Any] = tokenizer.encode_plus( A , A , add_special_tokens=A , return_special_tokens_mask=A , ) _UpperCAmelCase : Optional[int] = encoded_sequence_dict['''input_ids'''] _UpperCAmelCase : str = encoded_sequence_dict['''special_tokens_mask'''] self.assertEqual(len(A ) , len(A ) ) _UpperCAmelCase : Union[str, Any] = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(A ) ] _UpperCAmelCase : Optional[int] = [x for x in filtered_sequence if x is not None] self.assertEqual(A , A ) @require_tokenizers class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self ) -> int: # More context: # https://huggingface.co/wjmcat/opt-350m-paddle/discussions/1 # https://huggingface.slack.com/archives/C01N44FJDHT/p1653511495183519 # https://github.com/huggingface/transformers/pull/17088#discussion_r871246439 _UpperCAmelCase : Dict = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=A ) _UpperCAmelCase : Tuple = '''A photo of a cat''' _UpperCAmelCase : str = tokenizer.encode( A , ) self.assertEqual(A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''test_opt''' ) _UpperCAmelCase : Union[str, Any] = AutoTokenizer.from_pretrained('''./test_opt''' ) _UpperCAmelCase : Dict = tokenizer.encode( A , ) self.assertEqual(A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) def __lowerCAmelCase ( self ) -> Optional[Any]: _UpperCAmelCase : Dict = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , use_slow=A ) _UpperCAmelCase : Optional[Any] = '''A photo of a cat''' _UpperCAmelCase : Union[str, Any] = tokenizer.encode( A , ) # Same as above self.assertEqual(A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) @unittest.skip('''This test is failing because of a bug in the fast tokenizer''' ) def __lowerCAmelCase ( self ) -> Optional[Any]: _UpperCAmelCase : int = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=A ) _UpperCAmelCase : Optional[Any] = '''bos''' _UpperCAmelCase : Optional[int] = tokenizer.get_vocab()['''bos'''] _UpperCAmelCase : Tuple = '''A photo of a cat''' _UpperCAmelCase : Optional[Any] = tokenizer.encode( A , ) # We changed the bos token self.assertEqual(A , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''./tok''' ) _UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained('''./tok''' ) self.assertTrue(tokenizer.is_fast ) _UpperCAmelCase : List[str] = tokenizer.encode( A , ) self.assertEqual(A , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] )
263
"""simple docstring""" import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @require_torch def __lowerCAmelCase ( self ) -> Any: _UpperCAmelCase : str = pipeline( task='''zero-shot-audio-classification''' , model='''hf-internal-testing/tiny-clap-htsat-unfused''' ) _UpperCAmelCase : List[Any] = load_dataset('''ashraq/esc50''' ) _UpperCAmelCase : Optional[int] = dataset['''train''']['''audio'''][-1]['''array'''] _UpperCAmelCase : str = audio_classifier(A , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(A ) , [{'''score''': 0.501, '''label''': '''Sound of a dog'''}, {'''score''': 0.499, '''label''': '''Sound of vaccum cleaner'''}] , ) @unittest.skip('''No models are available in TF''' ) def __lowerCAmelCase ( self ) -> Union[str, Any]: pass @slow @require_torch def __lowerCAmelCase ( self ) -> str: _UpperCAmelCase : Union[str, Any] = pipeline( task='''zero-shot-audio-classification''' , model='''laion/clap-htsat-unfused''' , ) # This is an audio of a dog _UpperCAmelCase : List[Any] = load_dataset('''ashraq/esc50''' ) _UpperCAmelCase : Optional[int] = dataset['''train''']['''audio'''][-1]['''array'''] _UpperCAmelCase : Any = audio_classifier(A , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(A ) , [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ] , ) _UpperCAmelCase : List[Any] = audio_classifier([audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(A ) , [ [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) _UpperCAmelCase : Tuple = audio_classifier( [audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] , batch_size=5 ) self.assertEqual( nested_simplify(A ) , [ [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) @unittest.skip('''No models are available in TF''' ) def __lowerCAmelCase ( self ) -> int: pass
263
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Any = { """facebook/s2t-wav2vec2-large-en-de""": ( """https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json""" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase ( __lowerCamelCase ): UpperCAmelCase__ = 'speech_to_text_2' UpperCAmelCase__ = ['past_key_values'] UpperCAmelCase__ = {'num_attention_heads': 'decoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self : int , UpperCAmelCase : int=10000 , UpperCAmelCase : Optional[int]=6 , UpperCAmelCase : Union[str, Any]=2048 , UpperCAmelCase : Any=4 , UpperCAmelCase : str=0.0 , UpperCAmelCase : List[str]=True , UpperCAmelCase : Union[str, Any]="relu" , UpperCAmelCase : Union[str, Any]=256 , UpperCAmelCase : List[Any]=0.1 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : List[str]=0.0_2 , UpperCAmelCase : Dict=2 , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : List[str]=1 , UpperCAmelCase : Optional[Any]=0 , UpperCAmelCase : Tuple=2 , UpperCAmelCase : Tuple=1024 , **UpperCAmelCase : Union[str, Any] , ) -> int: lowerCamelCase__ : List[Any] = vocab_size lowerCamelCase__ : Dict = d_model lowerCamelCase__ : Any = decoder_ffn_dim lowerCamelCase__ : Optional[int] = decoder_layers lowerCamelCase__ : Any = decoder_attention_heads lowerCamelCase__ : Any = dropout lowerCamelCase__ : Any = attention_dropout lowerCamelCase__ : int = activation_dropout lowerCamelCase__ : Union[str, Any] = activation_function lowerCamelCase__ : Any = init_std lowerCamelCase__ : int = decoder_layerdrop lowerCamelCase__ : Tuple = use_cache lowerCamelCase__ : List[str] = decoder_layers lowerCamelCase__ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True lowerCamelCase__ : Tuple = max_target_positions super().__init__( pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , decoder_start_token_id=UpperCamelCase_ , **UpperCamelCase_ , )
357
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: if number > 0: raise ValueError('input must be a negative integer' ) lowerCamelCase__ : str = len(bin(_UpperCAmelCase )[3:] ) lowerCamelCase__ : Dict = bin(abs(_UpperCAmelCase ) - (1 << binary_number_length) )[3:] lowerCamelCase__ : Optional[int] = ( ( '1' + '0' * (binary_number_length - len(_UpperCAmelCase )) + twos_complement_number ) if number < 0 else '0' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
45
0
"""simple docstring""" import functools def lowercase ( __snake_case : list[int] , __snake_case : list[int] ): # Validation if not isinstance(__snake_case , __snake_case ) or not all(isinstance(__snake_case , __snake_case ) for day in days ): raise ValueError('''The parameter days should be a list of integers''' ) if len(__snake_case ) != 3 or not all(isinstance(__snake_case , __snake_case ) for cost in costs ): raise ValueError('''The parameter costs should be a list of three integers''' ) if len(__snake_case ) == 0: return 0 if min(__snake_case ) <= 0: raise ValueError('''All days elements should be greater than 0''' ) if max(__snake_case ) >= 3_6_6: raise ValueError('''All days elements should be less than 366''' ) lowercase_ : List[str] = set(__snake_case ) @functools.cache def dynamic_programming(__snake_case : int ) -> int: if index > 3_6_5: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 3_0 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
1
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : '''simple docstring''' def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Tuple=8 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Optional[int]=99 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : List[Any]=36 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : str=0.0 , UpperCAmelCase__ : int=512 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : Optional[Any]=0.02 , UpperCAmelCase__ : Dict=3 , UpperCAmelCase__ : Optional[Any]=4 , UpperCAmelCase__ : Tuple=None , ) ->Union[str, Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Any: '''simple docstring''' A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length]) A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) A__ = ids_tensor([self.batch_size] , self.num_choices) A__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Union[str, Any]: '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Any: '''simple docstring''' A__ = self.get_config() A__ = 300 return config def SCREAMING_SNAKE_CASE ( self : List[str]) ->List[Any]: '''simple docstring''' ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = self.prepare_config_and_inputs() A__ = True A__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) A__ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple) ->List[str]: '''simple docstring''' A__ = MraModel(config=UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__) A__ = model(UpperCAmelCase__ , token_type_ids=UpperCAmelCase__) A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] , ) ->List[Any]: '''simple docstring''' A__ = True A__ = MraModel(UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , encoder_attention_mask=UpperCAmelCase__ , ) A__ = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , ) A__ = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : str) ->int: '''simple docstring''' A__ = MraForMaskedLM(config=UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any) ->Any: '''simple docstring''' A__ = MraForQuestionAnswering(config=UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any) ->List[Any]: '''simple docstring''' A__ = self.num_labels A__ = MraForSequenceClassification(UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str]) ->str: '''simple docstring''' A__ = self.num_labels A__ = MraForTokenClassification(config=UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def SCREAMING_SNAKE_CASE ( self : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Any , UpperCAmelCase__ : str , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int]) ->Optional[int]: '''simple docstring''' A__ = self.num_choices A__ = MraForMultipleChoice(config=UpperCAmelCase__) model.to(UpperCAmelCase__) model.eval() A__ = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() A__ = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() A__ = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() A__ = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def SCREAMING_SNAKE_CASE ( self : List[Any]) ->int: '''simple docstring''' A__ = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = config_and_inputs A__ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = () def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Tuple: '''simple docstring''' A__ = MraModelTester(self) A__ = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37) def SCREAMING_SNAKE_CASE ( self : List[Any]) ->Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Tuple) ->Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : str) ->List[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ = type self.model_tester.create_and_check_model(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[Any]) ->List[str]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[str]) ->str: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->str: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Tuple) ->Any: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Any) ->List[str]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase__) @slow def SCREAMING_SNAKE_CASE ( self : str) ->Optional[int]: '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = MraModel.from_pretrained(UpperCAmelCase__) self.assertIsNotNone(UpperCAmelCase__) @unittest.skip(reason='''MRA does not output attentions''') def SCREAMING_SNAKE_CASE ( self : Tuple) ->Any: '''simple docstring''' return @require_torch class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE ( self : Dict) ->Tuple: '''simple docstring''' A__ = MraModel.from_pretrained('''uw-madison/mra-base-512-4''') A__ = torch.arange(256).unsqueeze(0) with torch.no_grad(): A__ = model(UpperCAmelCase__)[0] A__ = torch.Size((1, 256, 768)) self.assertEqual(output.shape , UpperCAmelCase__) A__ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase__ , atol=1e-4)) @slow def SCREAMING_SNAKE_CASE ( self : List[str]) ->Optional[int]: '''simple docstring''' A__ = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''') A__ = torch.arange(256).unsqueeze(0) with torch.no_grad(): A__ = model(UpperCAmelCase__)[0] A__ = 50_265 A__ = torch.Size((1, 256, vocab_size)) self.assertEqual(output.shape , UpperCAmelCase__) A__ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase__ , atol=1e-4)) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Optional[int]: '''simple docstring''' A__ = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''') A__ = torch.arange(4_096).unsqueeze(0) with torch.no_grad(): A__ = model(UpperCAmelCase__)[0] A__ = 50_265 A__ = torch.Size((1, 4_096, vocab_size)) self.assertEqual(output.shape , UpperCAmelCase__) A__ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase__ , atol=1e-4))
231
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration _lowerCamelCase : Optional[Any] = """facebook/wmt19-en-de""" _lowerCamelCase : Optional[Any] = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model _lowerCamelCase : int = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) _lowerCamelCase : Union[str, Any] = FSMTForConditionalGeneration(config) print(F'''num of params {tiny_model.num_parameters()}''') # Test _lowerCamelCase : int = tokenizer(["""Making tiny model"""], return_tensors="""pt""") _lowerCamelCase : int = tiny_model(**batch) print("""test output:""", len(outputs.logits[0])) # Save _lowerCamelCase : str = """tiny-wmt19-en-de""" tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'''Generated {mname_tiny}''') # Upload # transformers-cli upload tiny-wmt19-en-de
231
1
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) lowerCamelCase : Any = logging.getLogger(__name__) lowerCamelCase : List[str] = tf.data.AUTOTUNE def snake_case_ ( ): __lowercase : Dict = argparse.ArgumentParser(description="""Train a masked language model on TPU.""" ) parser.add_argument( """--pretrained_model_config""" , type=lowerCAmelCase_ , default="""roberta-base""" , help="""The model config to use. Note that we don't copy the model's weights, only the config!""" , ) parser.add_argument( """--tokenizer""" , type=lowerCAmelCase_ , default="""unigram-tokenizer-wikitext""" , help="""The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model's vocab size.""" , ) parser.add_argument( """--per_replica_batch_size""" , type=lowerCAmelCase_ , default=8 , help="""Batch size per TPU core.""" , ) parser.add_argument( """--no_tpu""" , action="""store_true""" , help="""If set, run on CPU and don't try to initialize a TPU. Useful for debugging on non-TPU instances.""" , ) parser.add_argument( """--tpu_name""" , type=lowerCAmelCase_ , help="""Name of TPU resource to initialize. Should be blank on Colab, and 'local' on TPU VMs.""" , default="""local""" , ) parser.add_argument( """--tpu_zone""" , type=lowerCAmelCase_ , help="""Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes.""" , ) parser.add_argument( """--gcp_project""" , type=lowerCAmelCase_ , help="""Google cloud project name. Only used for non-Colab TPU nodes.""" ) parser.add_argument( """--bfloat16""" , action="""store_true""" , help="""Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU.""" , ) parser.add_argument( """--train_dataset""" , type=lowerCAmelCase_ , help="""Path to training dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--shuffle_buffer_size""" , type=lowerCAmelCase_ , default=2**18 , help="""Size of the shuffle buffer (in samples)""" , ) parser.add_argument( """--eval_dataset""" , type=lowerCAmelCase_ , help="""Path to evaluation dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--num_epochs""" , type=lowerCAmelCase_ , default=1 , help="""Number of epochs to train for.""" , ) parser.add_argument( """--learning_rate""" , type=lowerCAmelCase_ , default=1e-4 , help="""Learning rate to use for training.""" , ) parser.add_argument( """--weight_decay_rate""" , type=lowerCAmelCase_ , default=1e-3 , help="""Weight decay rate to use for training.""" , ) parser.add_argument( """--max_length""" , type=lowerCAmelCase_ , default=512 , help="""Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py""" , ) parser.add_argument( """--mlm_probability""" , type=lowerCAmelCase_ , default=0.15 , help="""Fraction of tokens to mask during training.""" , ) parser.add_argument("""--output_dir""" , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help="""Path to save model checkpoints to.""" ) parser.add_argument("""--hub_model_id""" , type=lowerCAmelCase_ , help="""Model ID to upload to on the Hugging Face Hub.""" ) __lowercase : Any = parser.parse_args() return args def snake_case_ ( lowerCAmelCase_ : Any ): try: if args.tpu_name: __lowercase : Tuple = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: __lowercase : Dict = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( """Couldn't connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or """ """--gcp_project. When running on a TPU VM, use --tpu_name local.""" ) tf.config.experimental_connect_to_cluster(lowerCAmelCase_ ) tf.tpu.experimental.initialize_tpu_system(lowerCAmelCase_ ) return tpu def snake_case_ ( lowerCAmelCase_ : List[str] ): __lowercase : str = 0 for file in file_list: __lowercase : List[Any] = file.split("""/""" )[-1] __lowercase : Union[str, Any] = re.search(r"""-\d+-(\d+)\.tfrecord""" , lowerCAmelCase_ ).group(1 ) __lowercase : Any = int(lowerCAmelCase_ ) num_samples += sample_count return num_samples def snake_case_ ( lowerCAmelCase_ : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : List[str]=None ): __lowercase : str = count_samples(lowerCAmelCase_ ) __lowercase : List[str] = tf.data.Dataset.from_tensor_slices(lowerCAmelCase_ ) if shuffle: __lowercase : Optional[int] = dataset.shuffle(len(lowerCAmelCase_ ) ) __lowercase : List[str] = tf.data.TFRecordDataset(lowerCAmelCase_ , num_parallel_reads=lowerCAmelCase_ ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here __lowercase : Tuple = dataset.apply(tf.data.experimental.assert_cardinality(lowerCAmelCase_ ) ) __lowercase : Dict = dataset.map(lowerCAmelCase_ , num_parallel_calls=lowerCAmelCase_ ) if shuffle: assert shuffle_buffer_size is not None __lowercase : Tuple = dataset.shuffle(args.shuffle_buffer_size ) __lowercase : str = dataset.batch(lowerCAmelCase_ , drop_remainder=lowerCAmelCase_ ) __lowercase : Any = dataset.map(lowerCAmelCase_ , num_parallel_calls=lowerCAmelCase_ ) __lowercase : Tuple = dataset.prefetch(lowerCAmelCase_ ) return dataset def snake_case_ ( lowerCAmelCase_ : Any ): if not args.no_tpu: __lowercase : List[str] = initialize_tpu(lowerCAmelCase_ ) __lowercase : str = tf.distribute.TPUStrategy(lowerCAmelCase_ ) else: __lowercase : Optional[int] = tf.distribute.OneDeviceStrategy(device="""/gpu:0""" ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy("""mixed_bfloat16""" ) __lowercase : Any = AutoTokenizer.from_pretrained(args.tokenizer ) __lowercase : Union[str, Any] = AutoConfig.from_pretrained(args.pretrained_model_config ) __lowercase : Optional[int] = tokenizer.vocab_size __lowercase : Optional[Any] = tf.io.gfile.glob(os.path.join(args.train_dataset , """*.tfrecord""" ) ) if not training_records: raise ValueError(F"No .tfrecord files found in {args.train_dataset}." ) __lowercase : Optional[int] = tf.io.gfile.glob(os.path.join(args.eval_dataset , """*.tfrecord""" ) ) if not eval_records: raise ValueError(F"No .tfrecord files found in {args.eval_dataset}." ) __lowercase : Tuple = count_samples(lowerCAmelCase_ ) __lowercase : Any = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) __lowercase : str = steps_per_epoch * args.num_epochs with strategy.scope(): __lowercase : str = TFAutoModelForMaskedLM.from_config(lowerCAmelCase_ ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built __lowercase , __lowercase : Optional[int] = create_optimizer( num_train_steps=lowerCAmelCase_ , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=lowerCAmelCase_ , metrics=["""accuracy"""] ) def decode_fn(lowerCAmelCase_ : int ): __lowercase : Optional[Any] = { """input_ids""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), """attention_mask""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(lowerCAmelCase_ , lowerCAmelCase_ ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. __lowercase : Union[str, Any] = DataCollatorForLanguageModeling( tokenizer=lowerCAmelCase_ , mlm_probability=args.mlm_probability , mlm=lowerCAmelCase_ , return_tensors="""tf""" ) def mask_with_collator(lowerCAmelCase_ : Optional[int] ): # TF really needs an isin() function __lowercase : Any = ( ~tf.cast(batch["""attention_mask"""] , tf.bool ) | (batch["""input_ids"""] == tokenizer.cls_token_id) | (batch["""input_ids"""] == tokenizer.sep_token_id) ) __lowercase , __lowercase : Optional[int] = data_collator.tf_mask_tokens( batch["""input_ids"""] , vocab_size=len(lowerCAmelCase_ ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=lowerCAmelCase_ , ) return batch __lowercase : Optional[Any] = args.per_replica_batch_size * strategy.num_replicas_in_sync __lowercase : Tuple = prepare_dataset( lowerCAmelCase_ , decode_fn=lowerCAmelCase_ , mask_fn=lowerCAmelCase_ , batch_size=lowerCAmelCase_ , shuffle=lowerCAmelCase_ , shuffle_buffer_size=args.shuffle_buffer_size , ) __lowercase : List[Any] = prepare_dataset( lowerCAmelCase_ , decode_fn=lowerCAmelCase_ , mask_fn=lowerCAmelCase_ , batch_size=lowerCAmelCase_ , shuffle=lowerCAmelCase_ , ) __lowercase : Any = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=lowerCAmelCase_ ) ) model.fit( lowerCAmelCase_ , validation_data=lowerCAmelCase_ , epochs=args.num_epochs , callbacks=lowerCAmelCase_ , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": lowerCamelCase : List[str] = parse_args() main(args)
233
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def snake_case_ ( lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Any , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : int ): # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __lowercase : Optional[int] = TapasConfig.from_json_file(lowerCAmelCase_ ) # set absolute/relative position embeddings parameter __lowercase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __lowercase : Union[str, Any] = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WTQ": # run_task_main.py hparams __lowercase : List[Any] = 4 __lowercase : Union[str, Any] = True # hparam_utils.py hparams __lowercase : Any = 0.664_694 __lowercase : Tuple = 0.207_951 __lowercase : Dict = 0.121_194 __lowercase : List[str] = True __lowercase : str = True __lowercase : Dict = False __lowercase : Tuple = 0.0_352_513 __lowercase : List[Any] = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __lowercase : Optional[int] = 4 __lowercase : int = False # hparam_utils.py hparams __lowercase : Tuple = 36.4_519 __lowercase : str = 0.903_421 __lowercase : List[Any] = 222.088 __lowercase : Union[str, Any] = True __lowercase : Tuple = True __lowercase : Union[str, Any] = True __lowercase : Optional[Any] = 0.763_141 __lowercase : str = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "TABFACT": __lowercase : List[Any] = TapasForSequenceClassification(config=lowerCAmelCase_ ) elif task == "MLM": __lowercase : Optional[int] = TapasForMaskedLM(config=lowerCAmelCase_ ) elif task == "INTERMEDIATE_PRETRAINING": __lowercase : Dict = TapasModel(config=lowerCAmelCase_ ) else: raise ValueError(F"Task {task} not supported." ) print(F"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Save pytorch-model (weights and configuration) print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(lowerCAmelCase_ ) # Save tokenizer files print(F"Save tokenizer files to {pytorch_dump_path}" ) __lowercase : Any = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + """vocab.txt""" , model_max_length=512 ) tokenizer.save_pretrained(lowerCAmelCase_ ) print("""Used relative position embeddings:""" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase : int = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
233
1
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def _UpperCAmelCase (UpperCamelCase_ : List[Any] ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = SwinConfig(image_size=192 ) if "base" in model_name: _lowerCAmelCase : Tuple = 6 _lowerCAmelCase : int = 128 _lowerCAmelCase : Optional[Any] = (2, 2, 18, 2) _lowerCAmelCase : int = (4, 8, 16, 32) elif "large" in model_name: _lowerCAmelCase : Optional[Any] = 12 _lowerCAmelCase : Any = 192 _lowerCAmelCase : Union[str, Any] = (2, 2, 18, 2) _lowerCAmelCase : int = (6, 12, 24, 48) else: raise ValueError("""Model not supported, only supports base and large variants""" ) _lowerCAmelCase : Dict = window_size _lowerCAmelCase : Tuple = embed_dim _lowerCAmelCase : Union[str, Any] = depths _lowerCAmelCase : Tuple = num_heads return config def _UpperCAmelCase (UpperCamelCase_ : Optional[Any] ): '''simple docstring''' if "encoder.mask_token" in name: _lowerCAmelCase : Union[str, Any] = name.replace("""encoder.mask_token""" , """embeddings.mask_token""" ) if "encoder.patch_embed.proj" in name: _lowerCAmelCase : Optional[int] = name.replace("""encoder.patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "encoder.patch_embed.norm" in name: _lowerCAmelCase : Dict = name.replace("""encoder.patch_embed.norm""" , """embeddings.norm""" ) if "attn.proj" in name: _lowerCAmelCase : Optional[int] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _lowerCAmelCase : Optional[int] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _lowerCAmelCase : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _lowerCAmelCase : List[Any] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _lowerCAmelCase : Any = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _lowerCAmelCase : Any = name.replace("""mlp.fc2""" , """output.dense""" ) if name == "encoder.norm.weight": _lowerCAmelCase : Optional[Any] = """layernorm.weight""" if name == "encoder.norm.bias": _lowerCAmelCase : Dict = """layernorm.bias""" if "decoder" in name: pass else: _lowerCAmelCase : Optional[Any] = """swin.""" + name return name def _UpperCAmelCase (UpperCamelCase_ : Tuple , UpperCamelCase_ : Dict ): '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCAmelCase : Any = orig_state_dict.pop(UpperCamelCase_ ) if "attn_mask" in key: pass elif "qkv" in key: _lowerCAmelCase : Any = key.split(""".""" ) _lowerCAmelCase : Union[str, Any] = int(key_split[2] ) _lowerCAmelCase : List[str] = int(key_split[4] ) _lowerCAmelCase : Optional[int] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCAmelCase : Dict = val[:dim, :] _lowerCAmelCase : Optional[Any] = val[ dim : dim * 2, : ] _lowerCAmelCase : Union[str, Any] = val[-dim:, :] else: _lowerCAmelCase : int = val[ :dim ] _lowerCAmelCase : str = val[ dim : dim * 2 ] _lowerCAmelCase : int = val[ -dim: ] else: _lowerCAmelCase : Dict = val return orig_state_dict def _UpperCAmelCase (UpperCamelCase_ : Any , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Any ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = torch.load(UpperCamelCase_ , map_location="""cpu""" )["""model"""] _lowerCAmelCase : Union[str, Any] = get_swin_config(UpperCamelCase_ ) _lowerCAmelCase : str = SwinForMaskedImageModeling(UpperCamelCase_ ) model.eval() _lowerCAmelCase : int = convert_state_dict(UpperCamelCase_ , UpperCamelCase_ ) model.load_state_dict(UpperCamelCase_ ) _lowerCAmelCase : Tuple = """http://images.cocodataset.org/val2017/000000039769.jpg""" _lowerCAmelCase : List[str] = ViTImageProcessor(size={"""height""": 192, """width""": 192} ) _lowerCAmelCase : Tuple = Image.open(requests.get(UpperCamelCase_ , stream=UpperCamelCase_ ).raw ) _lowerCAmelCase : Any = image_processor(images=UpperCamelCase_ , return_tensors="""pt""" ) with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**UpperCamelCase_ ).logits print(outputs.keys() ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(UpperCamelCase_ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(UpperCamelCase_ ) if push_to_hub: print(F"Pushing model and image processor for {model_name} to hub" ) model.push_to_hub(F"microsoft/{model_name}" ) image_processor.push_to_hub(F"microsoft/{model_name}" ) if __name__ == "__main__": _lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="swin-base-simmim-window6-192", type=str, choices=["swin-base-simmim-window6-192", "swin-large-simmim-window12-192"], help="Name of the Swin SimMIM model you'd like to convert.", ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth", type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _lowerCamelCase : List[str] = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
159
import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __snake_case (_a , _a , unittest.TestCase ): lowerCAmelCase__ = IFPipeline lowerCAmelCase__ = TEXT_TO_IMAGE_PARAMS - {"width", "height", "latents"} lowerCAmelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"latents"} def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[str]: '''simple docstring''' return self._get_dummy_components() def SCREAMING_SNAKE_CASE ( self : Optional[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str]=0 ) -> Optional[Any]: '''simple docstring''' if str(_UpperCAmelCase ).startswith("""mps""" ): _lowerCAmelCase : Tuple = torch.manual_seed(_UpperCAmelCase ) else: _lowerCAmelCase : str = torch.Generator(device=_UpperCAmelCase ).manual_seed(_UpperCAmelCase ) _lowerCAmelCase : List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1 ) def SCREAMING_SNAKE_CASE ( self : int ) -> str: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[Any]: '''simple docstring''' self._test_save_load_local() def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[str]: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[int]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __snake_case (unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : str ) -> Any: '''simple docstring''' _lowerCAmelCase : str = IFPipeline.from_pretrained("""DeepFloyd/IF-I-XL-v1.0""" , variant="""fp16""" , torch_dtype=torch.floataa ) _lowerCAmelCase : Tuple = IFSuperResolutionPipeline.from_pretrained( """DeepFloyd/IF-II-L-v1.0""" , variant="""fp16""" , torch_dtype=torch.floataa , text_encoder=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("""cuda""" ) _lowerCAmelCase , _lowerCAmelCase : Tuple = pipe_a.encode_prompt("""anime turtle""" , device="""cuda""" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img _lowerCAmelCase : Optional[Any] = IFImgaImgPipeline(**pipe_a.components ) _lowerCAmelCase : Optional[int] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting _lowerCAmelCase : List[Any] = IFInpaintingPipeline(**pipe_a.components ) _lowerCAmelCase : Optional[Any] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Tuple , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : Any ) -> Any: '''simple docstring''' _start_torch_memory_measurement() _lowerCAmelCase : str = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : Optional[int] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , num_inference_steps=2 , generator=_UpperCAmelCase , output_type="""np""" , ) _lowerCAmelCase : List[str] = output.images[0] assert image.shape == (64, 64, 3) _lowerCAmelCase : Optional[int] = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 _lowerCAmelCase : str = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) # pipeline 2 _start_torch_memory_measurement() _lowerCAmelCase : str = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : Any = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : List[str] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , image=_UpperCAmelCase , generator=_UpperCAmelCase , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase : Optional[Any] = output.images[0] assert image.shape == (256, 256, 3) _lowerCAmelCase : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 _lowerCAmelCase : Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' _start_torch_memory_measurement() _lowerCAmelCase : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : Optional[Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : Optional[int] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , image=_UpperCAmelCase , num_inference_steps=2 , generator=_UpperCAmelCase , output_type="""np""" , ) _lowerCAmelCase : Union[str, Any] = output.images[0] assert image.shape == (64, 64, 3) _lowerCAmelCase : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 _lowerCAmelCase : List[str] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) # pipeline 2 _start_torch_memory_measurement() _lowerCAmelCase : str = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : List[str] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , image=_UpperCAmelCase , original_image=_UpperCAmelCase , generator=_UpperCAmelCase , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase : List[Any] = output.images[0] assert image.shape == (256, 256, 3) _lowerCAmelCase : List[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 _lowerCAmelCase : Optional[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict ) -> List[Any]: '''simple docstring''' _start_torch_memory_measurement() _lowerCAmelCase : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : Union[str, Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : Optional[int] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , image=_UpperCAmelCase , mask_image=_UpperCAmelCase , num_inference_steps=2 , generator=_UpperCAmelCase , output_type="""np""" , ) _lowerCAmelCase : List[str] = output.images[0] assert image.shape == (64, 64, 3) _lowerCAmelCase : Optional[int] = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 _lowerCAmelCase : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) # pipeline 2 _start_torch_memory_measurement() _lowerCAmelCase : Union[str, Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : Tuple = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(_UpperCAmelCase ) _lowerCAmelCase : Optional[int] = pipe_a( prompt_embeds=_UpperCAmelCase , negative_prompt_embeds=_UpperCAmelCase , image=_UpperCAmelCase , mask_image=_UpperCAmelCase , original_image=_UpperCAmelCase , generator=_UpperCAmelCase , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase : Dict = output.images[0] assert image.shape == (256, 256, 3) _lowerCAmelCase : Any = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 _lowerCAmelCase : str = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(_UpperCAmelCase , _UpperCAmelCase ) def _UpperCAmelCase (): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
159
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase ( A_ ): A__ : Union[str, Any] = ["image_processor", "tokenizer"] A__ : Tuple = "ChineseCLIPImageProcessor" A__ : Union[str, Any] = ("BertTokenizer", "BertTokenizerFast") def __init__(self : int , snake_case__ : Optional[int]=None , snake_case__ : int=None , **snake_case__ : Optional[int] ) -> Tuple: '''simple docstring''' snake_case : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , snake_case__ , ) snake_case : Optional[Any] = kwargs.pop("feature_extractor" ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(snake_case__ , snake_case__ ) snake_case : List[Any] = self.image_processor def __call__(self : Tuple , snake_case__ : Any=None , snake_case__ : Union[str, Any]=None , snake_case__ : List[str]=None , **snake_case__ : Tuple ) -> int: '''simple docstring''' if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: snake_case : List[str] = self.tokenizer(snake_case__ , return_tensors=snake_case__ , **snake_case__ ) if images is not None: snake_case : List[str] = self.image_processor(snake_case__ , return_tensors=snake_case__ , **snake_case__ ) if text is not None and images is not None: snake_case : Dict = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case__ ) , tensor_type=snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[Any] , *snake_case__ : Union[str, Any] , **snake_case__ : Any ) -> str: '''simple docstring''' return self.tokenizer.batch_decode(*snake_case__ , **snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] , *snake_case__ : Optional[Any] , **snake_case__ : Dict ) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*snake_case__ , **snake_case__ ) @property def _SCREAMING_SNAKE_CASE (self : List[str] ) -> List[Any]: '''simple docstring''' snake_case : int = self.tokenizer.model_input_names snake_case : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _SCREAMING_SNAKE_CASE (self : List[Any] ) -> Tuple: '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , snake_case__ , ) return self.image_processor_class
59
lowerCamelCase_ = frozenset( [ '''prompt''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) lowerCamelCase_ = frozenset(['''prompt''', '''negative_prompt''']) lowerCamelCase_ = frozenset([]) lowerCamelCase_ = frozenset(['''image''']) lowerCamelCase_ = frozenset( [ '''image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCamelCase_ = frozenset(['''image''']) lowerCamelCase_ = frozenset( [ '''prompt''', '''image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) lowerCamelCase_ = frozenset(['''prompt''', '''image''', '''negative_prompt''']) lowerCamelCase_ = frozenset( [ # Text guided image variation with an image mask '''prompt''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) lowerCamelCase_ = frozenset(['''prompt''', '''image''', '''mask_image''', '''negative_prompt''']) lowerCamelCase_ = frozenset( [ # image variation with an image mask '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCamelCase_ = frozenset(['''image''', '''mask_image''']) lowerCamelCase_ = frozenset( [ '''example_image''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCamelCase_ = frozenset(['''example_image''', '''image''', '''mask_image''']) lowerCamelCase_ = frozenset(['''class_labels''']) lowerCamelCase_ = frozenset(['''class_labels''']) lowerCamelCase_ = frozenset(['''batch_size''']) lowerCamelCase_ = frozenset([]) lowerCamelCase_ = frozenset(['''batch_size''']) lowerCamelCase_ = frozenset([]) lowerCamelCase_ = frozenset( [ '''prompt''', '''audio_length_in_s''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) lowerCamelCase_ = frozenset(['''prompt''', '''negative_prompt''']) lowerCamelCase_ = frozenset(['''input_tokens''']) lowerCamelCase_ = frozenset(['''input_tokens'''])
244
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={ "facebook/xglm-564M": "https://huggingface.co/facebook/xglm-564M/resolve/main/config.json", # See all XGLM models at https://huggingface.co/models?filter=xglm } class UpperCamelCase ( lowercase_ ): lowercase = 'xglm' lowercase = ['past_key_values'] lowercase = { 'num_attention_heads': 'attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'num_layers', } def __init__( self ,__UpperCamelCase=25_6008 ,__UpperCamelCase=2048 ,__UpperCamelCase=1024 ,__UpperCamelCase=4096 ,__UpperCamelCase=24 ,__UpperCamelCase=16 ,__UpperCamelCase="gelu" ,__UpperCamelCase=0.1 ,__UpperCamelCase=0.1 ,__UpperCamelCase=0.0 ,__UpperCamelCase=0.0 ,__UpperCamelCase=0.02 ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=2 ,__UpperCamelCase=1 ,__UpperCamelCase=0 ,__UpperCamelCase=2 ,**__UpperCamelCase ,) -> List[str]: '''simple docstring''' lowercase_ : Union[str, Any] = vocab_size lowercase_ : Optional[Any] = max_position_embeddings lowercase_ : Tuple = d_model lowercase_ : int = ffn_dim lowercase_ : Tuple = num_layers lowercase_ : Optional[int] = attention_heads lowercase_ : Dict = activation_function lowercase_ : Any = dropout lowercase_ : Dict = attention_dropout lowercase_ : Union[str, Any] = activation_dropout lowercase_ : Dict = layerdrop lowercase_ : Dict = init_std lowercase_ : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True lowercase_ : Dict = use_cache super().__init__( pad_token_id=__UpperCamelCase ,bos_token_id=__UpperCamelCase ,eos_token_id=__UpperCamelCase ,decoder_start_token_id=__UpperCamelCase ,**__UpperCamelCase ,)
351
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCamelCase ( unittest.TestCase ): def _UpperCAmelCase ( self ,__UpperCamelCase ) -> List[str]: '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] ,model_result['ss'] ): lowercase_ : Dict = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : int = 'sshleifer/tiny-gpt2' lowercase_ : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=__UpperCamelCase ,multi_process=__UpperCamelCase ,) lowercase_ : Union[str, Any] = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' lowercase_ : List[str] = 'sgugger/tiny-distilbert-classification' lowercase_ : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,only_pretrain_model=__UpperCamelCase ,) lowercase_ : int = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Any = 'sshleifer/tiny-gpt2' lowercase_ : Any = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,) lowercase_ : Optional[Any] = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Dict = 'sshleifer/tiny-gpt2' lowercase_ : Tuple = AutoConfig.from_pretrained(__UpperCamelCase ) lowercase_ : str = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=__UpperCamelCase ,multi_process=__UpperCamelCase ,) lowercase_ : str = TensorFlowBenchmark(__UpperCamelCase ,[config] ) lowercase_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' lowercase_ : Any = 'sshleifer/tiny-gpt2' lowercase_ : Any = AutoConfig.from_pretrained(__UpperCamelCase ) lowercase_ : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,) lowercase_ : int = TensorFlowBenchmark(__UpperCamelCase ,[config] ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : int = 'sshleifer/tiny-gpt2' lowercase_ : List[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,) lowercase_ : List[str] = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : List[str] = 'sshleifer/tiny-gpt2' lowercase_ : Optional[int] = AutoConfig.from_pretrained(__UpperCamelCase ) lowercase_ : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,) lowercase_ : str = TensorFlowBenchmark(__UpperCamelCase ,[config] ) lowercase_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : str = 'patrickvonplaten/t5-tiny-random' lowercase_ : int = AutoConfig.from_pretrained(__UpperCamelCase ) lowercase_ : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=__UpperCamelCase ,) lowercase_ : List[str] = TensorFlowBenchmark(__UpperCamelCase ,configs=[config] ) lowercase_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('GPU' ) ) == 0 ,'Cannot do xla on CPU.' ) def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : Optional[int] = 'sshleifer/tiny-gpt2' lowercase_ : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=__UpperCamelCase ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,use_xla=__UpperCamelCase ,multi_process=__UpperCamelCase ,) lowercase_ : Union[str, Any] = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : List[str] = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Any = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=__UpperCamelCase ,save_to_csv=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,inference_time_csv_file=os.path.join(__UpperCamelCase ,'inf_time.csv' ) ,inference_memory_csv_file=os.path.join(__UpperCamelCase ,'inf_mem.csv' ) ,env_info_csv_file=os.path.join(__UpperCamelCase ,'env.csv' ) ,multi_process=__UpperCamelCase ,) lowercase_ : List[str] = TensorFlowBenchmark(__UpperCamelCase ) benchmark.run() self.assertTrue(Path(os.path.join(__UpperCamelCase ,'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__UpperCamelCase ,'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__UpperCamelCase ,'env.csv' ) ).exists() ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : int = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(__UpperCamelCase ): self.assertTrue(hasattr(__UpperCamelCase ,'sequential' ) ) self.assertTrue(hasattr(__UpperCamelCase ,'cumulative' ) ) self.assertTrue(hasattr(__UpperCamelCase ,'current' ) ) self.assertTrue(hasattr(__UpperCamelCase ,'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=__UpperCamelCase ,sequence_lengths=[8] ,batch_sizes=[1] ,log_filename=os.path.join(__UpperCamelCase ,'log.txt' ) ,log_print=__UpperCamelCase ,trace_memory_line_by_line=__UpperCamelCase ,eager_mode=__UpperCamelCase ,multi_process=__UpperCamelCase ,) lowercase_ : Dict = TensorFlowBenchmark(__UpperCamelCase ) lowercase_ : Any = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(__UpperCamelCase ,'log.txt' ) ).exists() )
321
0
'''simple docstring''' from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def __UpperCAmelCase ( A : Dict , A : Any ) -> Tuple: UpperCAmelCase_ : List[Any] = k_size // 2 UpperCAmelCase_ , UpperCAmelCase_ : Any = mgrid[0 - center : k_size - center, 0 - center : k_size - center] UpperCAmelCase_ : str = 1 / (2 * pi * sigma) * exp(-(square(A ) + square(A )) / (2 * square(A )) ) return g def __UpperCAmelCase ( A : Optional[int] , A : Tuple , A : List[str] ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : int = image.shape[0], image.shape[1] # dst image height and width UpperCAmelCase_ : int = height - k_size + 1 UpperCAmelCase_ : Tuple = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows UpperCAmelCase_ : Optional[int] = zeros((dst_height * dst_width, k_size * k_size) ) UpperCAmelCase_ : Tuple = 0 for i, j in product(range(A ) , range(A ) ): UpperCAmelCase_ : List[Any] = ravel(image[i : i + k_size, j : j + k_size] ) UpperCAmelCase_ : int = window row += 1 # turn the kernel into shape(k*k, 1) UpperCAmelCase_ : Union[str, Any] = gen_gaussian_kernel(A , A ) UpperCAmelCase_ : Optional[int] = ravel(A ) # reshape and get the dst image UpperCAmelCase_ : Optional[Any] = dot(A , A ).reshape(A , A ).astype(A ) return dst if __name__ == "__main__": # read original image _UpperCamelCase : Any = imread(R'../image_data/lena.jpg') # turn image in gray scale value _UpperCamelCase : Optional[Any] = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size _UpperCamelCase : Optional[Any] = gaussian_filter(gray, 3, sigma=1) _UpperCamelCase : List[str] = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow('gaussian filter with 3x3 mask', gaussianaxa) imshow('gaussian filter with 5x5 mask', gaussianaxa) waitKey()
304
'''simple docstring''' import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def __UpperCAmelCase ( A : int , A : Any="shi-labs/oneformer_demo" ) -> Dict: with open(hf_hub_download(A , A , repo_type='''dataset''' ) , '''r''' ) as f: UpperCAmelCase_ : Union[str, Any] = json.load(A ) UpperCAmelCase_ : Optional[int] = {} UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : str = [] for key, info in class_info.items(): UpperCAmelCase_ : Tuple = info['''name'''] class_names.append(info['''name'''] ) if info["isthing"]: thing_ids.append(int(A ) ) UpperCAmelCase_ : Any = thing_ids UpperCAmelCase_ : Union[str, Any] = class_names return metadata class snake_case__ ( unittest.TestCase): def __init__( self : Any , _A : str , _A : Optional[int]=7 , _A : Tuple=3 , _A : Tuple=30 , _A : List[Any]=4_00 , _A : Tuple=None , _A : Optional[Any]=True , _A : Optional[Any]=True , _A : Any=[0.5, 0.5, 0.5] , _A : Any=[0.5, 0.5, 0.5] , _A : List[str]=10 , _A : Optional[int]=False , _A : Union[str, Any]=2_55 , _A : List[Any]="shi-labs/oneformer_demo" , _A : str="ade20k_panoptic.json" , _A : List[Any]=10 , ) -> Any: UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : Optional[Any] = batch_size UpperCAmelCase_ : Optional[Any] = num_channels UpperCAmelCase_ : Tuple = min_resolution UpperCAmelCase_ : Optional[int] = max_resolution UpperCAmelCase_ : Dict = do_resize UpperCAmelCase_ : Tuple = {'''shortest_edge''': 32, '''longest_edge''': 13_33} if size is None else size UpperCAmelCase_ : int = do_normalize UpperCAmelCase_ : List[Any] = image_mean UpperCAmelCase_ : Dict = image_std UpperCAmelCase_ : str = class_info_file UpperCAmelCase_ : Optional[Any] = prepare_metadata(_A , _A ) UpperCAmelCase_ : Tuple = num_text UpperCAmelCase_ : Union[str, Any] = repo_path # for the post_process_functions UpperCAmelCase_ : Any = 2 UpperCAmelCase_ : Dict = 10 UpperCAmelCase_ : int = 10 UpperCAmelCase_ : Optional[Any] = 3 UpperCAmelCase_ : str = 4 UpperCAmelCase_ : int = num_labels UpperCAmelCase_ : Union[str, Any] = do_reduce_labels UpperCAmelCase_ : str = ignore_index def A ( self : Dict ) -> List[Any]: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def A ( self : Any , _A : List[Any] , _A : List[str]=False ) -> Optional[Any]: if not batched: UpperCAmelCase_ : Any = image_inputs[0] if isinstance(_A , Image.Image ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = image.size else: UpperCAmelCase_ , UpperCAmelCase_ : int = image.shape[1], image.shape[2] if w < h: UpperCAmelCase_ : Union[str, Any] = int(self.size['''shortest_edge'''] * h / w ) UpperCAmelCase_ : int = self.size['''shortest_edge'''] elif w > h: UpperCAmelCase_ : List[Any] = self.size['''shortest_edge'''] UpperCAmelCase_ : Any = int(self.size['''shortest_edge'''] * w / h ) else: UpperCAmelCase_ : Dict = self.size['''shortest_edge'''] UpperCAmelCase_ : str = self.size['''shortest_edge'''] else: UpperCAmelCase_ : Dict = [] for image in image_inputs: UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase_ : int = max(_A , key=lambda _A : item[0] )[0] UpperCAmelCase_ : List[str] = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width def A ( self : Tuple ) -> str: return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class snake_case__ ( UpperCamelCase , unittest.TestCase): a_ = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string a_ = image_processing_class def A ( self : Optional[int] ) -> Any: UpperCAmelCase_ : int = OneFormerImageProcessorTester(self ) @property def A ( self : Any ) -> int: return self.image_processing_tester.prepare_image_processor_dict() def A ( self : Optional[Any] ) -> List[Any]: UpperCAmelCase_ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , '''image_mean''' ) ) self.assertTrue(hasattr(_A , '''image_std''' ) ) self.assertTrue(hasattr(_A , '''do_normalize''' ) ) self.assertTrue(hasattr(_A , '''do_resize''' ) ) self.assertTrue(hasattr(_A , '''size''' ) ) self.assertTrue(hasattr(_A , '''ignore_index''' ) ) self.assertTrue(hasattr(_A , '''class_info_file''' ) ) self.assertTrue(hasattr(_A , '''num_text''' ) ) self.assertTrue(hasattr(_A , '''repo_path''' ) ) self.assertTrue(hasattr(_A , '''metadata''' ) ) self.assertTrue(hasattr(_A , '''do_reduce_labels''' ) ) def A ( self : Dict ) -> Dict: pass def A ( self : Tuple ) -> Dict: # Initialize image_processor UpperCAmelCase_ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase_ : str = prepare_image_inputs(self.image_processing_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input UpperCAmelCase_ : str = image_processor(image_inputs[0] , ['''semantic'''] , return_tensors='''pt''' ).pixel_values UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = self.image_processing_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = self.image_processing_tester.get_expected_values(_A , batched=_A ) UpperCAmelCase_ : int = image_processor( _A , ['''semantic'''] * len(_A ) , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Tuple: # Initialize image_processor UpperCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase_ : Dict = prepare_image_inputs(self.image_processing_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input UpperCAmelCase_ : List[str] = image_processor(image_inputs[0] , ['''semantic'''] , return_tensors='''pt''' ).pixel_values UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.image_processing_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase_ , UpperCAmelCase_ : str = self.image_processing_tester.get_expected_values(_A , batched=_A ) UpperCAmelCase_ : Tuple = image_processor( _A , ['''semantic'''] * len(_A ) , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Dict ) -> Union[str, Any]: # Initialize image_processor UpperCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase_ : Dict = prepare_image_inputs(self.image_processing_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input UpperCAmelCase_ : int = image_processor(image_inputs[0] , ['''semantic'''] , return_tensors='''pt''' ).pixel_values UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.image_processing_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase_ , UpperCAmelCase_ : int = self.image_processing_tester.get_expected_values(_A , batched=_A ) UpperCAmelCase_ : Optional[int] = image_processor( _A , ['''semantic'''] * len(_A ) , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : int , _A : Any=False , _A : List[Any]=False , _A : Any="np" ) -> str: UpperCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # prepare image and target UpperCAmelCase_ : Tuple = self.image_processing_tester.num_labels UpperCAmelCase_ : int = None UpperCAmelCase_ : Union[str, Any] = None UpperCAmelCase_ : str = prepare_image_inputs(self.image_processing_tester , equal_resolution=_A ) if with_segmentation_maps: UpperCAmelCase_ : Any = num_labels if is_instance_map: UpperCAmelCase_ : Any = list(range(_A ) ) * 2 UpperCAmelCase_ : Optional[Any] = dict(enumerate(_A ) ) UpperCAmelCase_ : Dict = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": UpperCAmelCase_ : Dict = [Image.fromarray(_A ) for annotation in annotations] UpperCAmelCase_ : Tuple = image_processor( _A , ['''semantic'''] * len(_A ) , _A , return_tensors='''pt''' , instance_id_to_semantic_id=_A , pad_and_return_pixel_mask=_A , ) return inputs def A ( self : int ) -> str: pass def A ( self : Tuple ) -> Union[str, Any]: def common(_A : Optional[int]=False , _A : str=None ): UpperCAmelCase_ : List[str] = self.comm_get_image_processor_inputs( with_segmentation_maps=_A , is_instance_map=_A , segmentation_type=_A ) UpperCAmelCase_ : List[Any] = inputs['''mask_labels'''] UpperCAmelCase_ : Optional[Any] = inputs['''class_labels'''] UpperCAmelCase_ : int = inputs['''pixel_values'''] UpperCAmelCase_ : Tuple = inputs['''text_inputs'''] # check the batch_size for mask_label, class_label, text_input in zip(_A , _A , _A ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(_A ) , self.image_processing_tester.num_text ) common() common(is_instance_map=_A ) common(is_instance_map=_A , segmentation_type='''pil''' ) common(is_instance_map=_A , segmentation_type='''pil''' ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase_ : int = np.zeros((20, 50) ) UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ : Dict = 1 UpperCAmelCase_ : List[Any] = 1 UpperCAmelCase_ : List[Any] = binary_mask_to_rle(_A ) self.assertEqual(len(_A ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def A ( self : Any ) -> List[Any]: UpperCAmelCase_ : int = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='''ade20k_panoptic.json''' , num_text=self.image_processing_tester.num_text , repo_path='''shi-labs/oneformer_demo''' , ) UpperCAmelCase_ : Any = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase_ : Union[str, Any] = fature_extractor.post_process_semantic_segmentation(_A ) self.assertEqual(len(_A ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) UpperCAmelCase_ : List[str] = [(1, 4) for i in range(self.image_processing_tester.batch_size )] UpperCAmelCase_ : Any = fature_extractor.post_process_semantic_segmentation(_A , target_sizes=_A ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def A ( self : Optional[Any] ) -> Tuple: UpperCAmelCase_ : Any = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='''ade20k_panoptic.json''' , num_text=self.image_processing_tester.num_text , repo_path='''shi-labs/oneformer_demo''' , ) UpperCAmelCase_ : Dict = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase_ : List[Any] = image_processor.post_process_instance_segmentation(_A , threshold=0 ) self.assertTrue(len(_A ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('''segmentation''' in el ) self.assertTrue('''segments_info''' in el ) self.assertEqual(type(el['''segments_info'''] ) , _A ) self.assertEqual( el['''segmentation'''].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : Optional[Any] = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='''ade20k_panoptic.json''' , num_text=self.image_processing_tester.num_text , repo_path='''shi-labs/oneformer_demo''' , ) UpperCAmelCase_ : Tuple = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase_ : List[Any] = image_processor.post_process_panoptic_segmentation(_A , threshold=0 ) self.assertTrue(len(_A ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('''segmentation''' in el ) self.assertTrue('''segments_info''' in el ) self.assertEqual(type(el['''segments_info'''] ) , _A ) self.assertEqual( el['''segmentation'''].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
304
1
from __future__ import annotations class __SCREAMING_SNAKE_CASE : def __init__( self , __lowerCAmelCase=None ): UpperCamelCase__ = data UpperCamelCase__ = None def __repr__( self ): UpperCamelCase__ = [] UpperCamelCase__ = self while temp: string_rep.append(f"""{temp.data}""" ) UpperCamelCase__ = temp.next return "->".join(__lowerCAmelCase ) def _UpperCamelCase (a__ :list ): """simple docstring""" if not elements_list: raise Exception("""The Elements List is empty""" ) UpperCamelCase__ = UpperCamelCase__ = Node(elements_list[0] ) for i in range(1 , len(a__ ) ): UpperCamelCase__ = Node(elements_list[i] ) UpperCamelCase__ = current.next return head def _UpperCamelCase (a__ :Node ): """simple docstring""" if head_node is not None and isinstance(a__ , a__ ): print_reverse(head_node.next ) print(head_node.data ) def _UpperCamelCase (): """simple docstring""" from doctest import testmod testmod() UpperCamelCase__ = make_linked_list([14, 52, 14, 12, 43] ) print("""Linked List:""" ) print(a__ ) print("""Elements in Reverse:""" ) print_reverse(a__ ) if __name__ == "__main__": main()
87
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch UpperCamelCase__ = random.Random() def _UpperCamelCase (a__ :Any , a__ :Union[str, Any]=1.0 , a__ :Tuple=None , a__ :str=None ): """simple docstring""" if rng is None: UpperCamelCase__ = global_rng UpperCamelCase__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase=7 , __lowerCAmelCase=400 , __lowerCAmelCase=2000 , __lowerCAmelCase=10 , __lowerCAmelCase=160 , __lowerCAmelCase=8 , __lowerCAmelCase=0.0 , __lowerCAmelCase=4000 , __lowerCAmelCase=False , __lowerCAmelCase=True , ): UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = min_seq_length UpperCamelCase__ = max_seq_length UpperCamelCase__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCamelCase__ = padding_value UpperCamelCase__ = sampling_rate UpperCamelCase__ = return_attention_mask UpperCamelCase__ = do_normalize UpperCamelCase__ = feature_size UpperCamelCase__ = chunk_length UpperCamelCase__ = hop_length def _lowerCamelCase ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowerCamelCase ( self , __lowerCAmelCase=False , __lowerCAmelCase=False ): def _flatten(__lowerCAmelCase ): return list(itertools.chain(*__lowerCAmelCase ) ) if equal_length: UpperCamelCase__ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size UpperCamelCase__ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase__ = [np.asarray(__lowerCAmelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case : int = WhisperFeatureExtractor if is_speech_available() else None def _lowerCamelCase ( self ): UpperCamelCase__ = WhisperFeatureExtractionTester(self ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase__ = feat_extract_first.save_pretrained(__lowerCAmelCase )[0] check_json_file_has_correct_format(__lowerCAmelCase ) UpperCamelCase__ = self.feature_extraction_class.from_pretrained(__lowerCAmelCase ) UpperCamelCase__ = feat_extract_first.to_dict() UpperCamelCase__ = feat_extract_second.to_dict() UpperCamelCase__ = feat_extract_first.mel_filters UpperCamelCase__ = feat_extract_second.mel_filters self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase__ = os.path.join(__lowerCAmelCase , """feat_extract.json""" ) feat_extract_first.to_json_file(__lowerCAmelCase ) UpperCamelCase__ = self.feature_extraction_class.from_json_file(__lowerCAmelCase ) UpperCamelCase__ = feat_extract_first.to_dict() UpperCamelCase__ = feat_extract_second.to_dict() UpperCamelCase__ = feat_extract_first.mel_filters UpperCamelCase__ = feat_extract_second.mel_filters self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus UpperCamelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ = [np.asarray(__lowerCAmelCase ) for speech_input in speech_inputs] # Test feature size UpperCamelCase__ = feature_extractor(__lowerCAmelCase , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input UpperCamelCase__ = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features UpperCamelCase__ = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test batched UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. UpperCamelCase__ = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCamelCase__ = np.asarray(__lowerCAmelCase ) UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test truncation required UpperCamelCase__ = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] UpperCamelCase__ = [np.asarray(__lowerCAmelCase ) for speech_input in speech_inputs] UpperCamelCase__ = [x[: feature_extractor.n_samples] for x in speech_inputs] UpperCamelCase__ = [np.asarray(__lowerCAmelCase ) for speech_input in speech_inputs_truncated] UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) def _lowerCamelCase ( self ): import torch UpperCamelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ = np.random.rand(100 , 32 ).astype(np.floataa ) UpperCamelCase__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCamelCase__ = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) UpperCamelCase__ = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _lowerCamelCase ( self , __lowerCAmelCase ): UpperCamelCase__ = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech UpperCamelCase__ = ds.sort("""id""" ).select(range(__lowerCAmelCase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _lowerCamelCase ( self ): # fmt: off UpperCamelCase__ = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on UpperCamelCase__ = self._load_datasamples(1 ) UpperCamelCase__ = WhisperFeatureExtractor() UpperCamelCase__ = feature_extractor(__lowerCAmelCase , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , __lowerCAmelCase , atol=1E-4 ) ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ = self._load_datasamples(1 )[0] UpperCamelCase__ = ((audio - audio.min()) / (audio.max() - audio.min())) * 65535 # Rescale to [0, 65535] to show issue UpperCamelCase__ = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=__lowerCAmelCase )[0] self.assertTrue(np.all(np.mean(__lowerCAmelCase ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__lowerCAmelCase ) - 1 ) < 1E-3 ) )
87
1
from __future__ import annotations import bisect def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ = 0,snake_case_ = -1 ): if hi < 0: _A : List[Any] = len(snake_case_ ) while lo < hi: _A : str = lo + (hi - lo) // 2 if sorted_collection[mid] < item: _A : str = mid + 1 else: _A : Tuple = mid return lo def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ = 0,snake_case_ = -1 ): if hi < 0: _A : List[str] = len(snake_case_ ) while lo < hi: _A : Tuple = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: _A : Union[str, Any] = mid + 1 else: _A : Any = mid return lo def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ = 0,snake_case_ = -1 ): sorted_collection.insert(bisect_left(snake_case_,snake_case_,snake_case_,snake_case_ ),snake_case_ ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ = 0,snake_case_ = -1 ): sorted_collection.insert(bisect_right(snake_case_,snake_case_,snake_case_,snake_case_ ),snake_case_ ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : Optional[int] = 0 _A : Optional[Any] = len(snake_case_ ) - 1 while left <= right: _A : List[str] = left + (right - left) // 2 _A : List[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: _A : List[Any] = midpoint - 1 else: _A : List[str] = midpoint + 1 return None def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : int = bisect.bisect_left(snake_case_,snake_case_ ) if index != len(snake_case_ ) and sorted_collection[index] == item: return index return None def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_ ): if right < left: return None _A : List[str] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(snake_case_,snake_case_,snake_case_,midpoint - 1 ) else: return binary_search_by_recursion(snake_case_,snake_case_,midpoint + 1,snake_case_ ) if __name__ == "__main__": _snake_case = input("Enter numbers separated by comma:\n").strip() _snake_case = sorted(int(item) for item in user_input.split(",")) _snake_case = int(input("Enter a single number to be found in the list:\n")) _snake_case = binary_search(collection, target) if result is None: print(f"""{target} was not found in {collection}.""") else: print(f"""{target} was found at position {result} in {collection}.""")
26
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : Optional[int] = logging.get_logger(__name__) _lowerCAmelCase : List[str] = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'git_vision_model' def __init__( self , __snake_case=768 , __snake_case=3072 , __snake_case=12 , __snake_case=12 , __snake_case=3 , __snake_case=224 , __snake_case=16 , __snake_case="quick_gelu" , __snake_case=1e-5 , __snake_case=0.0 , __snake_case=0.02 , **__snake_case , ) -> int: '''simple docstring''' super().__init__(**__snake_case ) __a =hidden_size __a =intermediate_size __a =num_hidden_layers __a =num_attention_heads __a =num_channels __a =patch_size __a =image_size __a =initializer_range __a =attention_dropout __a =layer_norm_eps __a =hidden_act @classmethod def __magic_name__ ( cls , __snake_case , **__snake_case ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(__snake_case ) __a , __a =cls.get_config_dict(__snake_case , **__snake_case ) # get the vision config dict if we are loading from GITConfig if config_dict.get('model_type' ) == "git": __a =config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(__snake_case , **__snake_case ) class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'git' def __init__( self , __snake_case=None , __snake_case=3_0522 , __snake_case=768 , __snake_case=6 , __snake_case=12 , __snake_case=3072 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=1024 , __snake_case=0.02 , __snake_case=1e-12 , __snake_case=0 , __snake_case="absolute" , __snake_case=True , __snake_case=False , __snake_case=101 , __snake_case=102 , __snake_case=None , **__snake_case , ) -> Optional[int]: '''simple docstring''' super().__init__(bos_token_id=__snake_case , eos_token_id=__snake_case , pad_token_id=__snake_case , **__snake_case ) if vision_config is None: __a ={} logger.info('vision_config is None. initializing the GitVisionConfig with default values.' ) __a =GitVisionConfig(**__snake_case ) __a =vocab_size __a =hidden_size __a =num_hidden_layers __a =num_attention_heads __a =hidden_act __a =intermediate_size __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =max_position_embeddings __a =initializer_range __a =layer_norm_eps __a =position_embedding_type __a =use_cache __a =tie_word_embeddings __a =num_image_with_embedding __a =bos_token_id __a =eos_token_id def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' __a =copy.deepcopy(self.__dict__ ) __a =self.vision_config.to_dict() __a =self.__class__.model_type return output
218
0
import numpy as np from PIL import Image def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): __UpperCamelCase =np.array(SCREAMING_SNAKE_CASE__ ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) __UpperCamelCase =0 __UpperCamelCase =0 __UpperCamelCase =0 __UpperCamelCase =0 # compute the shape of the output matrix __UpperCamelCase =(arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __UpperCamelCase =np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __UpperCamelCase =np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCamelCase =0 __UpperCamelCase =0 return updated_arr def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): __UpperCamelCase =np.array(SCREAMING_SNAKE_CASE__ ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) __UpperCamelCase =0 __UpperCamelCase =0 __UpperCamelCase =0 __UpperCamelCase =0 # compute the shape of the output matrix __UpperCamelCase =(arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __UpperCamelCase =np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __UpperCamelCase =int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCamelCase =0 __UpperCamelCase =0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='avgpooling', verbose=True) # Loading the image _A = Image.open('path_to_image') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
117
from ....utils import logging _A = logging.get_logger(__name__) class UpperCAmelCase__ ( A_ ): """simple docstring""" def __init__( self , A_ , A_=None , A_=2048 ) -> Any: __UpperCamelCase =config.__dict__ __UpperCamelCase =modal_hidden_size if num_labels: __UpperCamelCase =num_labels
117
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __A = { "configuration_convbert": ["CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvBertConfig", "ConvBertOnnxConfig"], "tokenization_convbert": ["ConvBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["ConvBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvBertForMaskedLM", "ConvBertForMultipleChoice", "ConvBertForQuestionAnswering", "ConvBertForSequenceClassification", "ConvBertForTokenClassification", "ConvBertLayer", "ConvBertModel", "ConvBertPreTrainedModel", "load_tf_weights_in_convbert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFConvBertForMaskedLM", "TFConvBertForMultipleChoice", "TFConvBertForQuestionAnswering", "TFConvBertForSequenceClassification", "TFConvBertForTokenClassification", "TFConvBertLayer", "TFConvBertModel", "TFConvBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
148
from heapq import heappop, heappush import numpy as np def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) -> tuple[float | int, list[tuple[int, int]]]: '''simple docstring''' lowercase , lowercase : Optional[int] = grid.shape lowercase : Optional[int] = [-1, 1, 0, 0] lowercase : List[str] = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] lowercase , lowercase : Union[str, Any] = [(0, source)], set() lowercase : List[str] = np.full((rows, cols) , np.inf ) lowercase : Dict = 0 lowercase : Dict = np.empty((rows, cols) , dtype=__magic_name__ ) lowercase : Any = None while queue: ((lowercase) , (lowercase)) : Optional[Any] = heappop(__magic_name__ ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: lowercase : Tuple = [] while (x, y) != source: path.append((x, y) ) lowercase , lowercase : Optional[int] = predecessors[x, y] path.append(__magic_name__ ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(__magic_name__ ) ): lowercase , lowercase : Optional[int] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: lowercase : List[Any] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(__magic_name__ , (dist + 1, (nx, ny)) ) lowercase : int = dist + 1 lowercase : Optional[Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
308
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class _UpperCamelCase ( A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = XLMRobertaTokenizer lowerCAmelCase__ = XLMRobertaTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def __lowerCamelCase ( self : List[Any]): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowercase =XLMRobertaTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase) tokenizer.save_pretrained(self.tmpdirname) def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase ='<pad>' __lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase) , _lowerCAmelCase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase) , _lowerCAmelCase) def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase =list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(_lowerCAmelCase) , 1_0_0_2) def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2) def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =XLMRobertaTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase) __lowercase =tokenizer.tokenize('This is a test') self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est']) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) __lowercase =tokenizer.tokenize('I was born in 92000, and this is falsé.') self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __lowercase =tokenizer.convert_tokens_to_ids(_lowerCAmelCase) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __lowercase =tokenizer.convert_ids_to_tokens(_lowerCAmelCase) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __lowercase =(self.rust_tokenizer_class, 'hf-internal-testing/tiny-xlm-roberta', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})"""): __lowercase =self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase) __lowercase =self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase) __lowercase =tempfile.mkdtemp() __lowercase =tokenizer_r.save_pretrained(_lowerCAmelCase) __lowercase =tokenizer_p.save_pretrained(_lowerCAmelCase) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files)) __lowercase =tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f) self.assertSequenceEqual(_lowerCAmelCase , _lowerCAmelCase) # Checks everything loads correctly in the same way __lowercase =tokenizer_r.from_pretrained(_lowerCAmelCase) __lowercase =tokenizer_p.from_pretrained(_lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_lowerCAmelCase) # Save tokenizer rust, legacy_format=True __lowercase =tempfile.mkdtemp() __lowercase =tokenizer_r.save_pretrained(_lowerCAmelCase , legacy_format=_lowerCAmelCase) __lowercase =tokenizer_p.save_pretrained(_lowerCAmelCase) # Checks it save with the same files self.assertSequenceEqual(_lowerCAmelCase , _lowerCAmelCase) # Checks everything loads correctly in the same way __lowercase =tokenizer_r.from_pretrained(_lowerCAmelCase) __lowercase =tokenizer_p.from_pretrained(_lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase)) shutil.rmtree(_lowerCAmelCase) # Save tokenizer rust, legacy_format=False __lowercase =tempfile.mkdtemp() __lowercase =tokenizer_r.save_pretrained(_lowerCAmelCase , legacy_format=_lowerCAmelCase) __lowercase =tokenizer_p.save_pretrained(_lowerCAmelCase) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __lowercase =tokenizer_r.from_pretrained(_lowerCAmelCase) __lowercase =tokenizer_p.from_pretrained(_lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase)) shutil.rmtree(_lowerCAmelCase) @cached_property def __lowerCamelCase ( self : Tuple): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained('xlm-roberta-base') def __lowerCamelCase ( self : Tuple): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(_lowerCAmelCase , f.name) __lowercase =XLMRobertaTokenizer(f.name , keep_accents=_lowerCAmelCase) __lowercase =pickle.dumps(_lowerCAmelCase) pickle.loads(_lowerCAmelCase) def __lowerCamelCase ( self : List[str]): '''simple docstring''' if not self.test_rust_tokenizer: return __lowercase =self.get_tokenizer() __lowercase =self.get_rust_tokenizer() __lowercase ='I was born in 92000, and this is falsé.' __lowercase =tokenizer.tokenize(_lowerCAmelCase) __lowercase =rust_tokenizer.tokenize(_lowerCAmelCase) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase) __lowercase =tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase) __lowercase =rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase) __lowercase =self.get_rust_tokenizer() __lowercase =tokenizer.encode(_lowerCAmelCase) __lowercase =rust_tokenizer.encode(_lowerCAmelCase) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase) @slow def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase ='Hello World!' __lowercase =[0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase)) @slow def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase =( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) __lowercase =[ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase)) @slow def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' __lowercase ={'input_ids': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='xlm-roberta-base' , revision='d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3' , )
48
'''simple docstring''' import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class _UpperCamelCase ( A ): '''simple docstring''' lowerCAmelCase__ = """bart""" lowerCAmelCase__ = ["""past_key_values"""] lowerCAmelCase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , _lowerCAmelCase : Any=5_0_2_6_5 , _lowerCAmelCase : Optional[Any]=1_0_2_4 , _lowerCAmelCase : List[Any]=1_2 , _lowerCAmelCase : Any=4_0_9_6 , _lowerCAmelCase : List[str]=1_6 , _lowerCAmelCase : List[Any]=1_2 , _lowerCAmelCase : Dict=4_0_9_6 , _lowerCAmelCase : Optional[Any]=1_6 , _lowerCAmelCase : List[Any]=0.0 , _lowerCAmelCase : Optional[Any]=0.0 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : str=1_0_2_4 , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Union[str, Any]=0.0 , _lowerCAmelCase : List[str]=0.0 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : List[str]=False , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : Dict=0 , _lowerCAmelCase : int=2 , _lowerCAmelCase : Any=True , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : str=2 , **_lowerCAmelCase : Optional[int] , ): '''simple docstring''' __lowercase =vocab_size __lowercase =max_position_embeddings __lowercase =d_model __lowercase =encoder_ffn_dim __lowercase =encoder_layers __lowercase =encoder_attention_heads __lowercase =decoder_ffn_dim __lowercase =decoder_layers __lowercase =decoder_attention_heads __lowercase =dropout __lowercase =attention_dropout __lowercase =activation_dropout __lowercase =activation_function __lowercase =init_std __lowercase =encoder_layerdrop __lowercase =decoder_layerdrop __lowercase =classifier_dropout __lowercase =use_cache __lowercase =encoder_layers __lowercase =scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=_lowerCAmelCase , pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , is_encoder_decoder=_lowerCAmelCase , decoder_start_token_id=_lowerCAmelCase , forced_eos_token_id=_lowerCAmelCase , **_lowerCAmelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , _lowerCAmelCase): __lowercase =self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ 'The config can simply be saved and uploaded again to be fixed.') class _UpperCamelCase ( A ): '''simple docstring''' @property def __lowerCamelCase ( self : List[Any]): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __lowercase =OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ]) if self.use_past: __lowercase ={0: 'batch'} __lowercase ={0: 'batch', 1: 'past_decoder_sequence + sequence'} else: __lowercase ={0: 'batch', 1: 'decoder_sequence'} __lowercase ={0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(_lowerCAmelCase , direction='inputs') elif self.task == "causal-lm": # TODO: figure this case out. __lowercase =OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ]) if self.use_past: __lowercase , __lowercase =self.num_layers for i in range(_lowerCAmelCase): __lowercase ={0: 'batch', 2: 'past_sequence + sequence'} __lowercase ={0: 'batch', 2: 'past_sequence + sequence'} else: __lowercase =OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ]) return common_inputs @property def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __lowercase =super().outputs else: __lowercase =super(_lowerCAmelCase , self).outputs if self.use_past: __lowercase , __lowercase =self.num_layers for i in range(_lowerCAmelCase): __lowercase ={0: 'batch', 2: 'past_sequence + sequence'} __lowercase ={0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def __lowerCamelCase ( self : Tuple , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : int = -1 , _lowerCAmelCase : int = -1 , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[TensorType] = None , ): '''simple docstring''' __lowercase =self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) # Generate decoder inputs __lowercase =seq_length if not self.use_past else 1 __lowercase =self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) __lowercase ={f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} __lowercase =dict(**_lowerCAmelCase , **_lowerCAmelCase) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.') else: import torch __lowercase , __lowercase =common_inputs['input_ids'].shape __lowercase =common_inputs['decoder_input_ids'].shape[1] __lowercase , __lowercase =self.num_attention_heads __lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase =decoder_seq_length + 3 __lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __lowercase =torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(_lowerCAmelCase , _lowerCAmelCase)] , dim=1) __lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered __lowercase , __lowercase =self.num_layers __lowercase =min(_lowerCAmelCase , _lowerCAmelCase) __lowercase =max(_lowerCAmelCase , _lowerCAmelCase) - min_num_layers __lowercase ='encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(_lowerCAmelCase): common_inputs["past_key_values"].append( ( torch.zeros(_lowerCAmelCase), torch.zeros(_lowerCAmelCase), torch.zeros(_lowerCAmelCase), torch.zeros(_lowerCAmelCase), )) # TODO: test this. __lowercase =encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(_lowerCAmelCase , _lowerCAmelCase): common_inputs["past_key_values"].append((torch.zeros(_lowerCAmelCase), torch.zeros(_lowerCAmelCase))) return common_inputs def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : int = -1 , _lowerCAmelCase : int = -1 , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[TensorType] = None , ): '''simple docstring''' __lowercase =self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.') else: import torch __lowercase , __lowercase =common_inputs['input_ids'].shape # Not using the same length for past_key_values __lowercase =seqlen + 2 __lowercase , __lowercase =self.num_layers __lowercase , __lowercase =self.num_attention_heads __lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase =common_inputs['attention_mask'].dtype __lowercase =torch.cat( [common_inputs['attention_mask'], torch.ones(_lowerCAmelCase , _lowerCAmelCase , dtype=_lowerCAmelCase)] , dim=1) __lowercase =[ (torch.zeros(_lowerCAmelCase), torch.zeros(_lowerCAmelCase)) for _ in range(_lowerCAmelCase) ] return common_inputs def __lowerCamelCase ( self : Optional[int] , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : int = -1 , _lowerCAmelCase : int = -1 , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[TensorType] = None , ): '''simple docstring''' __lowercase =compute_effective_axis_dimension( _lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __lowercase =tokenizer.num_special_tokens_to_add(_lowerCAmelCase) __lowercase =compute_effective_axis_dimension( _lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_lowerCAmelCase) # Generate dummy inputs according to compute batch and sequence __lowercase =[' '.join([tokenizer.unk_token]) * seq_length] * batch_size __lowercase =dict(tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase)) return common_inputs def __lowerCamelCase ( self : Optional[Any] , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : int = -1 , _lowerCAmelCase : int = -1 , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[TensorType] = None , ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( _lowerCAmelCase , batch_size=_lowerCAmelCase , seq_length=_lowerCAmelCase , is_pair=_lowerCAmelCase , framework=_lowerCAmelCase) elif self.task == "causal-lm": __lowercase =self._generate_dummy_inputs_for_causal_lm( _lowerCAmelCase , batch_size=_lowerCAmelCase , seq_length=_lowerCAmelCase , is_pair=_lowerCAmelCase , framework=_lowerCAmelCase) else: __lowercase =self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCAmelCase , batch_size=_lowerCAmelCase , seq_length=_lowerCAmelCase , is_pair=_lowerCAmelCase , framework=_lowerCAmelCase) return common_inputs def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any]): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __lowercase =super()._flatten_past_key_values_(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) else: __lowercase =super(_lowerCAmelCase , self)._flatten_past_key_values_( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase)
48
1
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase_ : Dict = logging.getLogger(__name__) UpperCAmelCase_ : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) UpperCAmelCase_ : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _SCREAMING_SNAKE_CASE : snake_case__ : Optional[str] = field( default=_a , metadata={ """help""": ( """The model checkpoint for weights initialization. Leave None if you want to train a model from""" """ scratch.""" ) } , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(_a )} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class _SCREAMING_SNAKE_CASE : snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """The input training data file (a text file)."""} ) snake_case__ : Optional[str] = field( default=_a , metadata={ """help""": ( """The input training data files (multiple files in glob format). """ """Very often splitting large files to smaller files can prevent tokenizer going out of memory""" ) } , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input train ref data file for whole word mask in Chinese."""} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input eval ref data file for whole word mask in Chinese."""} , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Whether distinct lines of text in the dataset are to be handled as distinct sequences."""} , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Train with masked-language modeling loss instead of language modeling."""} ) snake_case__ : bool = field(default=_a , metadata={"""help""": """Whether ot not to use whole word mask."""} ) snake_case__ : float = field( default=0.15 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) snake_case__ : float = field( default=1 / 6 , metadata={ """help""": ( """Ratio of length of a span of masked tokens to surrounding context length for permutation language""" """ modeling.""" ) } , ) snake_case__ : int = field( default=5 , metadata={"""help""": """Maximum length of a span of masked tokens for permutation language modeling."""} ) snake_case__ : int = field( default=-1 , metadata={ """help""": ( """Optional input sequence length after tokenization.""" """The training dataset will be truncated in block of this size for training.""" """Default to the model max input length for single sentence inputs (take into account special tokens).""" ) } , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : DataTrainingArguments , __magic_name__ : PreTrainedTokenizer , __magic_name__ : bool = False , __magic_name__ : Optional[str] = None , ) -> List[Any]: """simple docstring""" def _dataset(__magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any]=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError("""You need to set world whole masking and mlm to True for Chinese Whole Word Mask""" ) return LineByLineWithRefDataset( tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size , ref_path=__magic_name__ , ) return LineByLineTextDataset(tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size ) else: return TextDataset( tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=__magic_name__ , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(__magic_name__ ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def SCREAMING_SNAKE_CASE_ ( ) -> Optional[int]: """simple docstring""" UpperCamelCase :List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCamelCase , UpperCamelCase , UpperCamelCase :Dict = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( """Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file """ """or remove the --do_eval argument.""" ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , __magic_name__ ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: UpperCamelCase :Tuple = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase :Optional[Any] = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: UpperCamelCase :Dict = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.tokenizer_name: UpperCamelCase :Tuple = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase :int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( """You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another""" """ script, save it,and load it from here, using --tokenizer_name""" ) if model_args.model_name_or_path: UpperCamelCase :Dict = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) else: logger.info("""Training new model from scratch""" ) UpperCamelCase :Any = AutoModelWithLMHead.from_config(__magic_name__ ) model.resize_token_embeddings(len(__magic_name__ ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( """BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the""" """--mlm flag (masked language modeling).""" ) if data_args.block_size <= 0: UpperCamelCase :Dict = tokenizer.max_len # Our input block size will be the max possible for the model else: UpperCamelCase :int = min(data_args.block_size , tokenizer.max_len ) # Get datasets UpperCamelCase :Optional[int] = ( get_dataset(__magic_name__ , tokenizer=__magic_name__ , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) UpperCamelCase :Any = ( get_dataset(__magic_name__ , tokenizer=__magic_name__ , evaluate=__magic_name__ , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": UpperCamelCase :int = DataCollatorForPermutationLanguageModeling( tokenizer=__magic_name__ , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: UpperCamelCase :str = DataCollatorForWholeWordMask( tokenizer=__magic_name__ , mlm_probability=data_args.mlm_probability ) else: UpperCamelCase :Any = DataCollatorForLanguageModeling( tokenizer=__magic_name__ , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer UpperCamelCase :List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , data_collator=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , prediction_loss_only=__magic_name__ , ) # Training if training_args.do_train: UpperCamelCase :Optional[Any] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=__magic_name__ ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase :Optional[int] = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCamelCase :List[Any] = trainer.evaluate() UpperCamelCase :List[Any] = math.exp(eval_output["""eval_loss"""] ) UpperCamelCase :Optional[Any] = {"""perplexity""": perplexity} UpperCamelCase :Tuple = os.path.join(training_args.output_dir , """eval_results_lm.txt""" ) if trainer.is_world_master(): with open(__magic_name__ , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key in sorted(result.keys() ): logger.info(""" %s = %s""" , __magic_name__ , str(result[key] ) ) writer.write("""%s = %s\n""" % (key, str(result[key] )) ) results.update(__magic_name__ ) return results def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Union[str, Any] ) -> int: """simple docstring""" main() if __name__ == "__main__": main()
38
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase_ ( __UpperCAmelCase : Any , __UpperCAmelCase : str=False ) -> List[str]: SCREAMING_SNAKE_CASE_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"module.blocks.{i}.norm1.weight", f"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((f"module.blocks.{i}.norm1.bias", f"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (f"module.blocks.{i}.attn.proj.weight", f"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((f"module.blocks.{i}.attn.proj.bias", f"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((f"module.blocks.{i}.norm2.weight", f"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((f"module.blocks.{i}.norm2.bias", f"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((f"module.blocks.{i}.mlp.fc1.weight", f"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((f"module.blocks.{i}.mlp.fc1.bias", f"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((f"module.blocks.{i}.mlp.fc2.weight", f"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((f"module.blocks.{i}.mlp.fc2.bias", f"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ('module.cls_token', 'vit.embeddings.cls_token'), ('module.patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('module.patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('module.pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('module.norm.weight', 'layernorm.weight'), ('module.norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" SCREAMING_SNAKE_CASE_ = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def UpperCAmelCase_ ( __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any]=False ) -> Any: for i in range(config.num_hidden_layers ): if base_model: SCREAMING_SNAKE_CASE_ = '' else: SCREAMING_SNAKE_CASE_ = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE_ = state_dict.pop(f"module.blocks.{i}.attn.qkv.weight" ) SCREAMING_SNAKE_CASE_ = state_dict.pop(f"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE_ = in_proj_weight[ : config.hidden_size, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE_ = in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[-config.hidden_size :] def UpperCAmelCase_ ( __UpperCAmelCase : List[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : Optional[int] ) -> str: # projection head is used in the self-supervised pre-training in MSN, # for downstream task it's not needed. SCREAMING_SNAKE_CASE_ = [ 'module.fc.fc1.weight', 'module.fc.fc1.bias', 'module.fc.bn1.weight', 'module.fc.bn1.bias', 'module.fc.bn1.running_mean', 'module.fc.bn1.running_var', 'module.fc.bn1.num_batches_tracked', 'module.fc.fc2.weight', 'module.fc.fc2.bias', 'module.fc.bn2.weight', 'module.fc.bn2.bias', 'module.fc.bn2.running_mean', 'module.fc.bn2.running_var', 'module.fc.bn2.num_batches_tracked', 'module.fc.fc3.weight', 'module.fc.fc3.bias', ] for k in ignore_keys: state_dict.pop(__UpperCAmelCase , __UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : Dict , __UpperCAmelCase : Any , __UpperCAmelCase : Any ) -> Dict: SCREAMING_SNAKE_CASE_ = dct.pop(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = val def UpperCAmelCase_ ( __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple ) -> List[str]: SCREAMING_SNAKE_CASE_ = ViTMSNConfig() SCREAMING_SNAKE_CASE_ = 10_00 SCREAMING_SNAKE_CASE_ = 'datasets/huggingface/label-files' SCREAMING_SNAKE_CASE_ = 'imagenet-1k-id2label.json' SCREAMING_SNAKE_CASE_ = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase ) , 'r' ) ) SCREAMING_SNAKE_CASE_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ = idalabel SCREAMING_SNAKE_CASE_ = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: SCREAMING_SNAKE_CASE_ = 3_84 SCREAMING_SNAKE_CASE_ = 15_36 SCREAMING_SNAKE_CASE_ = 6 elif "l16" in checkpoint_url: SCREAMING_SNAKE_CASE_ = 10_24 SCREAMING_SNAKE_CASE_ = 40_96 SCREAMING_SNAKE_CASE_ = 24 SCREAMING_SNAKE_CASE_ = 16 SCREAMING_SNAKE_CASE_ = 0.1 elif "b4" in checkpoint_url: SCREAMING_SNAKE_CASE_ = 4 elif "l7" in checkpoint_url: SCREAMING_SNAKE_CASE_ = 7 SCREAMING_SNAKE_CASE_ = 10_24 SCREAMING_SNAKE_CASE_ = 40_96 SCREAMING_SNAKE_CASE_ = 24 SCREAMING_SNAKE_CASE_ = 16 SCREAMING_SNAKE_CASE_ = 0.1 SCREAMING_SNAKE_CASE_ = ViTMSNModel(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = torch.hub.load_state_dict_from_url(__UpperCAmelCase , map_location='cpu' )['target_encoder'] SCREAMING_SNAKE_CASE_ = ViTImageProcessor(size=config.image_size ) remove_projection_head(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = create_rename_keys(__UpperCAmelCase , base_model=__UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , base_model=__UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE_ = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) SCREAMING_SNAKE_CASE_ = ViTImageProcessor( size=config.image_size , image_mean=__UpperCAmelCase , image_std=__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = image_processor(images=__UpperCAmelCase , return_tensors='pt' ) # forward pass torch.manual_seed(2 ) SCREAMING_SNAKE_CASE_ = model(**__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: SCREAMING_SNAKE_CASE_ = torch.tensor([[-1.0_9_1_5, -1.4_8_7_6, -1.1_8_0_9]] ) elif "b16" in checkpoint_url: SCREAMING_SNAKE_CASE_ = torch.tensor([[1_4.2_8_8_9, -1_8.9_0_4_5, 1_1.7_2_8_1]] ) elif "l16" in checkpoint_url: SCREAMING_SNAKE_CASE_ = torch.tensor([[4_1.5_0_2_8, -2_2.8_6_8_1, 4_5.6_4_7_5]] ) elif "b4" in checkpoint_url: SCREAMING_SNAKE_CASE_ = torch.tensor([[-4.3_8_6_8, 5.2_9_3_2, -0.4_1_3_7]] ) else: SCREAMING_SNAKE_CASE_ = torch.tensor([[-0.1_7_9_2, -0.6_4_6_5, 2.4_2_6_3]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , __UpperCAmelCase , atol=1E-4 ) print(f"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__UpperCAmelCase ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) lowerCamelCase__ : Union[str, Any] = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
225
0
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor snake_case : Any = logging.get_logger(__name__) class _snake_case ( _snake_case ): def __init__( self , *_lowerCamelCase , **_lowerCamelCase ): warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , _lowerCamelCase , ) super().__init__(*_lowerCamelCase , **_lowerCamelCase )
281
from ...configuration_utils import PretrainedConfig class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = 'bert-generation' def __init__( self , _lowerCamelCase=5_0358 , _lowerCamelCase=1024 , _lowerCamelCase=24 , _lowerCamelCase=16 , _lowerCamelCase=4096 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=0.02 , _lowerCamelCase=1e-12 , _lowerCamelCase=0 , _lowerCamelCase=2 , _lowerCamelCase=1 , _lowerCamelCase="absolute" , _lowerCamelCase=True , **_lowerCamelCase , ): super().__init__(pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , **_lowerCamelCase ) a :Optional[int] = vocab_size a :Tuple = hidden_size a :Any = num_hidden_layers a :Any = num_attention_heads a :List[Any] = hidden_act a :Tuple = intermediate_size a :Any = hidden_dropout_prob a :int = attention_probs_dropout_prob a :Dict = max_position_embeddings a :int = initializer_range a :Union[str, Any] = layer_norm_eps a :str = position_embedding_type a :int = use_cache
281
1
'''simple docstring''' import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _lowerCamelCase : Dict = get_logger(__name__) class SCREAMING_SNAKE_CASE ( enum.Enum ): """simple docstring""" _SCREAMING_SNAKE_CASE = """all_checks""" _SCREAMING_SNAKE_CASE = """basic_checks""" _SCREAMING_SNAKE_CASE = """no_checks""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __lowerCamelCase ( A__ , A__ , A__=None ) -> List[Any]: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(A__ ) - set(A__ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(A__ ) - set(A__ ) ) ) if len(set(A__ ) - set(A__ ) ) > 0: raise UnexpectedDownloadedFile(str(set(A__ ) - set(A__ ) ) ) UpperCamelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] UpperCamelCase = ' for ' + verification_name if verification_name is not None else '' if len(A__ ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __lowerCamelCase ( A__ , A__ ) -> Tuple: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(A__ ) - set(A__ ) ) > 0: raise ExpectedMoreSplits(str(set(A__ ) - set(A__ ) ) ) if len(set(A__ ) - set(A__ ) ) > 0: raise UnexpectedSplits(str(set(A__ ) - set(A__ ) ) ) UpperCamelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(A__ ) > 0: raise NonMatchingSplitsSizesError(str(A__ ) ) logger.info('All the splits matched successfully.' ) def __lowerCamelCase ( A__ , A__ = True ) -> dict: """simple docstring""" if record_checksum: UpperCamelCase = shaaaa() with open(A__ , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'' ): m.update(A__ ) UpperCamelCase = m.hexdigest() else: UpperCamelCase = None return {"num_bytes": os.path.getsize(A__ ), "checksum": checksum} def __lowerCamelCase ( A__ ) -> Optional[int]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
28
"""simple docstring""" def lowercase ( lowerCAmelCase__ : str , lowerCAmelCase__ : list[str] ) -> str: __a = '''''' for word_or_phrase in separated: if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise Exception('''join() accepts only strings to be joined''' ) joined += word_or_phrase + separator return joined.strip(lowerCAmelCase__ ) if __name__ == "__main__": from doctest import testmod testmod()
45
0
'''simple docstring''' def snake_case__ ( lowerCamelCase__ : int , lowerCamelCase__ : int ) -> str: if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) A_ : Optional[int] = str(bin(_lowercase ) )[2:] # remove the leading "0b" A_ : Optional[Any] = str(bin(_lowercase ) )[2:] # remove the leading "0b" A_ : str = max(len(_lowercase ) , len(_lowercase ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(_lowercase ) , b_binary.zfill(_lowercase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
362
'''simple docstring''' from collections.abc import Sequence def snake_case__ ( lowerCamelCase__ : Sequence[float] , lowerCamelCase__ : bool = False ) -> float: if not arr: return 0 A_ : Union[str, Any] = 0 if allow_empty_subarrays else float('''-inf''' ) A_ : str = 0.0 for num in arr: A_ : Any = max(0 if allow_empty_subarrays else num , curr_sum + num ) A_ : Tuple = max(lowerCamelCase__ , lowerCamelCase__ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() snake_case__ = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'{max_subarray_sum(nums) = }')
4
0
import sys def lowerCamelCase__ ( __lowerCAmelCase : Optional[Any] ): """simple docstring""" lowerCAmelCase_ = len(__lowerCAmelCase ) lowerCAmelCase_ = [[0 for x in range(__lowerCAmelCase )] for x in range(__lowerCAmelCase )] lowerCAmelCase_ = [[0 for x in range(__lowerCAmelCase )] for x in range(__lowerCAmelCase )] for chain_length in range(2 , __lowerCAmelCase ): for a in range(1 , n - chain_length + 1 ): lowerCAmelCase_ = a + chain_length - 1 lowerCAmelCase_ = sys.maxsize for c in range(__lowerCAmelCase , __lowerCAmelCase ): lowerCAmelCase_ = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: lowerCAmelCase_ = cost lowerCAmelCase_ = c return matrix, sol def lowerCamelCase__ ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : int ): """simple docstring""" if i == j: print("A" + str(__lowerCAmelCase ) , end=" " ) else: print("(" , end=" " ) print_optiomal_solution(__lowerCAmelCase , __lowerCAmelCase , optimal_solution[i][j] ) print_optiomal_solution(__lowerCAmelCase , optimal_solution[i][j] + 1 , __lowerCAmelCase ) print(")" , end=" " ) def lowerCamelCase__ ( ): """simple docstring""" lowerCAmelCase_ = [30, 35, 15, 5, 10, 20, 25] lowerCAmelCase_ = len(__lowerCAmelCase ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 lowerCAmelCase_ , lowerCAmelCase_ = matrix_chain_order(__lowerCAmelCase ) print("No. of Operation required: " + str(matrix[1][n - 1] ) ) print_optiomal_solution(__lowerCAmelCase , 1 , n - 1 ) if __name__ == "__main__": main()
231
import os import re import shutil import sys import tempfile import unittest import black _A = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. _A = " \"\"\"\n Output class for the scheduler's step function output.\n\n Args:\n prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the\n denoising loop.\n pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n The predicted denoised sample (x_{0}) based on the model output from the current timestep.\n `pred_original_sample` can be used to preview progress or for guidance.\n \"\"\"\n\n prev_sample: torch.FloatTensor\n pred_original_sample: Optional[torch.FloatTensor] = None\n" class _lowerCAmelCase ( unittest.TestCase ): def __a ( self ) -> str: lowerCAmelCase_ = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir , "schedulers/" ) ) lowerCAmelCase_ = self.diffusers_dir shutil.copy( os.path.join(_UpperCamelCase , "src/diffusers/schedulers/scheduling_ddpm.py" ) , os.path.join(self.diffusers_dir , "schedulers/scheduling_ddpm.py" ) , ) def __a ( self ) -> Optional[int]: lowerCAmelCase_ = "src/diffusers" shutil.rmtree(self.diffusers_dir ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None ) -> Optional[Any]: lowerCAmelCase_ = comment + f"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: lowerCAmelCase_ = comment + f"""\nclass {class_name}(nn.Module):\n""" + overwrite_result lowerCAmelCase_ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) lowerCAmelCase_ = black.format_str(_UpperCamelCase , mode=_UpperCamelCase ) lowerCAmelCase_ = os.path.join(self.diffusers_dir , "new_code.py" ) with open(_UpperCamelCase , "w" , newline="\n" ) as f: f.write(_UpperCamelCase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(_UpperCamelCase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=_UpperCamelCase ) with open(_UpperCamelCase , "r" ) as f: self.assertTrue(f.read() , _UpperCamelCase ) def __a ( self ) -> Optional[Any]: lowerCAmelCase_ = check_copies.find_code_in_diffusers("schedulers.scheduling_ddpm.DDPMSchedulerOutput" ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def __a ( self ) -> Tuple: # Base copy consistency self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput" , "DDPMSchedulerOutput" , REFERENCE_CODE + "\n" , ) # With no empty line at the end self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput" , "DDPMSchedulerOutput" , _UpperCamelCase , ) # Copy consistency with rename self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test" , "TestSchedulerOutput" , re.sub("DDPM" , "Test" , _UpperCamelCase ) , ) # Copy consistency with a really long name lowerCAmelCase_ = "TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason" self.check_copy_consistency( f"""# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}""" , f"""{long_class_name}SchedulerOutput""" , re.sub("Bert" , _UpperCamelCase , _UpperCamelCase ) , ) # Copy consistency with overwrite self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test" , "TestSchedulerOutput" , _UpperCamelCase , overwrite_result=re.sub("DDPM" , "Test" , _UpperCamelCase ) , )
231
1
from __future__ import annotations from collections.abc import MutableSequence class UpperCamelCase : '''simple docstring''' def __init__( self , UpperCamelCase_ , UpperCamelCase_ ): if len(_a ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) lowercase_ :list[float] = list(_a ) lowercase_ :Dict = degree def __add__( self , UpperCamelCase_ ): if self.degree > polynomial_a.degree: lowercase_ :Optional[int] = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _a ) else: lowercase_ :List[Any] = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _a ) def __sub__( self , UpperCamelCase_ ): return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self ): return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self , UpperCamelCase_ ): lowercase_ :list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _a ) def UpperCamelCase ( self , UpperCamelCase_ ): lowercase_ :int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): lowercase_ :Optional[int] = "" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_a ) return polynomial def __repr__( self ): return self.__str__() def UpperCamelCase ( self ): lowercase_ :list[float] = [0] * self.degree for i in range(self.degree ): lowercase_ :List[Any] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _a ) def UpperCamelCase ( self , UpperCamelCase_ = 0 ): lowercase_ :list[float] = [0] * (self.degree + 2) lowercase_ :List[str] = constant for i in range(self.degree + 1 ): lowercase_ :Tuple = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _a ) def __eq__( self , UpperCamelCase_ ): if not isinstance(_a , _a ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self , UpperCamelCase_ ): return not self.__eq__(_a )
370
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self ): lowercase_ :List[Any] = 1 lowercase_ :List[Any] = 3 lowercase_ :str = (32, 32) lowercase_ :Tuple = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCamelCase_ ) return image @property def UpperCamelCase ( self ): torch.manual_seed(0 ) lowercase_ :Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def UpperCamelCase ( self ): torch.manual_seed(0 ) lowercase_ :str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def UpperCamelCase ( self ): torch.manual_seed(0 ) lowercase_ :int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModel(UpperCamelCase_ ) @property def UpperCamelCase ( self ): def extract(*UpperCamelCase_ , **UpperCamelCase_ ): class UpperCamelCase : '''simple docstring''' def __init__( self ): lowercase_ :Dict = torch.ones([0] ) def UpperCamelCase ( self , UpperCamelCase_ ): self.pixel_values.to(UpperCamelCase_ ) return self return Out() return extract def UpperCamelCase ( self ): lowercase_ :Optional[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowercase_ :List[Any] = self.dummy_cond_unet lowercase_ :int = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=UpperCamelCase_ , set_alpha_to_one=UpperCamelCase_ , ) lowercase_ :Any = self.dummy_vae lowercase_ :Dict = self.dummy_text_encoder lowercase_ :str = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # make sure here that pndm scheduler skips prk lowercase_ :Optional[Any] = StableDiffusionPipeline( unet=UpperCamelCase_ , scheduler=UpperCamelCase_ , vae=UpperCamelCase_ , text_encoder=UpperCamelCase_ , tokenizer=UpperCamelCase_ , safety_checker=UpperCamelCase_ , feature_extractor=self.dummy_extractor , ) lowercase_ :Union[str, Any] = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :str = '''A painting of a squirrel eating a burger''' lowercase_ :int = torch.Generator(device=UpperCamelCase_ ).manual_seed(0 ) lowercase_ :Optional[Any] = sd_pipe([prompt] , generator=UpperCamelCase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' ) lowercase_ :Any = output.images lowercase_ :List[Any] = torch.Generator(device=UpperCamelCase_ ).manual_seed(0 ) lowercase_ :List[Any] = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=UpperCamelCase_ , )[0] lowercase_ :Dict = image[0, -3:, -3:, -1] lowercase_ :List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ :List[Any] = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self ): lowercase_ :List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowercase_ :List[str] = self.dummy_cond_unet lowercase_ :Optional[Any] = PNDMScheduler(skip_prk_steps=UpperCamelCase_ ) lowercase_ :Optional[Any] = self.dummy_vae lowercase_ :List[Any] = self.dummy_text_encoder lowercase_ :str = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # make sure here that pndm scheduler skips prk lowercase_ :Tuple = StableDiffusionPipeline( unet=UpperCamelCase_ , scheduler=UpperCamelCase_ , vae=UpperCamelCase_ , text_encoder=UpperCamelCase_ , tokenizer=UpperCamelCase_ , safety_checker=UpperCamelCase_ , feature_extractor=self.dummy_extractor , ) lowercase_ :Optional[int] = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :str = '''A painting of a squirrel eating a burger''' lowercase_ :Any = torch.Generator(device=UpperCamelCase_ ).manual_seed(0 ) lowercase_ :Optional[int] = sd_pipe([prompt] , generator=UpperCamelCase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' ) lowercase_ :Optional[Any] = output.images lowercase_ :List[Any] = torch.Generator(device=UpperCamelCase_ ).manual_seed(0 ) lowercase_ :List[str] = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=UpperCamelCase_ , )[0] lowercase_ :Dict = image[0, -3:, -3:, -1] lowercase_ :str = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ :Dict = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self ): lowercase_ :List[str] = StableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-lms-pipe''' , safety_checker=UpperCamelCase_ ) assert isinstance(UpperCamelCase_ , UpperCamelCase_ ) assert isinstance(pipe.scheduler , UpperCamelCase_ ) assert pipe.safety_checker is None lowercase_ :Optional[int] = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCamelCase_ ) lowercase_ :Union[str, Any] = StableDiffusionPipeline.from_pretrained(UpperCamelCase_ ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowercase_ :List[Any] = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def UpperCamelCase ( self ): lowercase_ :Optional[Any] = self.dummy_cond_unet lowercase_ :Any = PNDMScheduler(skip_prk_steps=UpperCamelCase_ ) lowercase_ :int = self.dummy_vae lowercase_ :Tuple = self.dummy_text_encoder lowercase_ :Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # put models in fp16 lowercase_ :Optional[int] = unet.half() lowercase_ :Union[str, Any] = vae.half() lowercase_ :Optional[int] = bert.half() # make sure here that pndm scheduler skips prk lowercase_ :Any = StableDiffusionPipeline( unet=UpperCamelCase_ , scheduler=UpperCamelCase_ , vae=UpperCamelCase_ , text_encoder=UpperCamelCase_ , tokenizer=UpperCamelCase_ , safety_checker=UpperCamelCase_ , feature_extractor=self.dummy_extractor , ) lowercase_ :Dict = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :List[str] = '''A painting of a squirrel eating a burger''' lowercase_ :List[Any] = sd_pipe([prompt] , num_inference_steps=2 , output_type='''np''' ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self ): lowercase_ :Optional[Any] = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=UpperCamelCase_ ) lowercase_ :Dict = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase_ :List[Any] = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :List[Any] = ( '''portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle''' ''' coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with''' ''' anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and''' ''' children from bahnhof zoo, detailed ''' ) lowercase_ :str = 40_0366_0346 lowercase_ :Optional[Any] = 7 # without safety guidance (sld_guidance_scale = 0) lowercase_ :Tuple = torch.manual_seed(UpperCamelCase_ ) lowercase_ :int = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) lowercase_ :List[str] = output.images lowercase_ :int = image[0, -3:, -3:, -1] lowercase_ :str = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 # without safety guidance (strong configuration) lowercase_ :Dict = torch.manual_seed(UpperCamelCase_ ) lowercase_ :Any = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ :int = output.images lowercase_ :Union[str, Any] = image[0, -3:, -3:, -1] lowercase_ :Optional[int] = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self ): lowercase_ :Tuple = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=UpperCamelCase_ ) lowercase_ :List[Any] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase_ :Dict = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :Optional[int] = '''padme amidala taking a bath artwork, safe for work, no nudity''' lowercase_ :Any = 27_3497_1755 lowercase_ :str = 7 lowercase_ :Optional[Any] = torch.manual_seed(UpperCamelCase_ ) lowercase_ :Tuple = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) lowercase_ :Optional[Any] = output.images lowercase_ :str = image[0, -3:, -3:, -1] lowercase_ :int = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 lowercase_ :Any = torch.manual_seed(UpperCamelCase_ ) lowercase_ :List[Any] = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ :List[str] = output.images lowercase_ :Optional[Any] = image[0, -3:, -3:, -1] lowercase_ :Optional[Any] = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self ): lowercase_ :Tuple = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' ) lowercase_ :Tuple = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase_ :List[str] = ( '''the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.''' ''' leyendecker''' ) lowercase_ :Any = 10_4435_5234 lowercase_ :Union[str, Any] = 12 lowercase_ :str = torch.manual_seed(UpperCamelCase_ ) lowercase_ :str = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) lowercase_ :Optional[int] = output.images lowercase_ :str = image[0, -3:, -3:, -1] lowercase_ :Optional[int] = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-7 lowercase_ :Dict = torch.manual_seed(UpperCamelCase_ ) lowercase_ :Optional[Any] = sd_pipe( [prompt] , generator=UpperCamelCase_ , guidance_scale=UpperCamelCase_ , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ :Optional[Any] = output.images lowercase_ :List[Any] = image[0, -3:, -3:, -1] lowercase_ :Any = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
252
0
import math class __lowerCAmelCase : """simple docstring""" def __init__( self : Any , _lowerCAmelCase : Tuple=0 ) -> str: # a graph with Node 0,1,...,N-1 """simple docstring""" snake_case_ = n snake_case_ = [ [math.inf for j in range(0 , _lowerCAmelCase )] for i in range(0 , _lowerCAmelCase ) ] # adjacency matrix for weight snake_case_ = [ [math.inf for j in range(0 , _lowerCAmelCase )] for i in range(0 , _lowerCAmelCase ) ] # dp[i][j] stores minimum distance from i to j def lowerCAmelCase__ ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : List[str] ) -> Optional[int]: """simple docstring""" snake_case_ = w def lowerCAmelCase__ ( self : str ) -> Optional[Any]: """simple docstring""" for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): snake_case_ = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def lowerCAmelCase__ ( self : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict ) -> Optional[int]: """simple docstring""" return self.dp[u][v] if __name__ == "__main__": SCREAMING_SNAKE_CASE :Tuple = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
159
import unittest from knapsack import greedy_knapsack as kp class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase__ ( self : Any ) -> str: """simple docstring""" snake_case_ = [1_0, 2_0, 3_0, 4_0, 5_0, 6_0] snake_case_ = [2, 4, 6, 8, 1_0, 1_2] snake_case_ = 1_0_0 self.assertEqual(kp.calc_profit(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , 2_1_0 ) def lowerCAmelCase__ ( self : List[str] ) -> Tuple: """simple docstring""" self.assertRaisesRegex(_lowerCAmelCase , "max_weight must greater than zero." ) def lowerCAmelCase__ ( self : Dict ) -> Optional[Any]: """simple docstring""" self.assertRaisesRegex(_lowerCAmelCase , "Weight can not be negative." ) def lowerCAmelCase__ ( self : Dict ) -> Tuple: """simple docstring""" self.assertRaisesRegex(_lowerCAmelCase , "Profit can not be negative." ) def lowerCAmelCase__ ( self : Optional[int] ) -> str: """simple docstring""" self.assertRaisesRegex(_lowerCAmelCase , "max_weight must greater than zero." ) def lowerCAmelCase__ ( self : Union[str, Any] ) -> str: """simple docstring""" self.assertRaisesRegex( _lowerCAmelCase , "The length of profit and weight must be same." ) if __name__ == "__main__": unittest.main()
159
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = CodeGenTokenizer a_ = CodeGenTokenizerFast a_ = True a_ = {'add_prefix_space': True} a_ = False def _lowercase ( self : Any ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case__ : Dict = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] snake_case__ : Dict = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE ) ) ) ) snake_case__ : str = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] snake_case__ : str = {"unk_token": "<unk>"} snake_case__ : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) snake_case__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_SCREAMING_SNAKE_CASE ) ) def _lowercase ( self : Optional[Any] , **__A : Union[str, Any] ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE ) def _lowercase ( self : Optional[Any] , **__A : Tuple ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE ) def _lowercase ( self : Dict , __A : str ): snake_case__ : Optional[int] = "lower newer" snake_case__ : Dict = "lower newer" return input_text, output_text def _lowercase ( self : Union[str, Any] ): snake_case__ : str = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case__ : Any = "lower newer" snake_case__ : Any = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] snake_case__ : Optional[Any] = tokenizer.tokenize(_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = tokens + [tokenizer.unk_token] snake_case__ : List[Any] = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) def _lowercase ( self : List[str] ): if not self.test_rust_tokenizer: return snake_case__ : Union[str, Any] = self.get_tokenizer() snake_case__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=_SCREAMING_SNAKE_CASE ) snake_case__ : Any = "lower newer" # Testing tokenization snake_case__ : List[str] = tokenizer.tokenize(_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = rust_tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Testing conversion to ids without special tokens snake_case__ : Optional[int] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE ) snake_case__ : str = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Testing conversion to ids with special tokens snake_case__ : Union[str, Any] = self.get_rust_tokenizer(add_prefix_space=_SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE ) snake_case__ : Tuple = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Testing the unknown token snake_case__ : int = tokens + [rust_tokenizer.unk_token] snake_case__ : Union[str, Any] = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) def _lowercase ( self : Union[str, Any] , *__A : str , **__A : Dict ): # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def _lowercase ( self : str , __A : Dict=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): snake_case__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # Simple input snake_case__ : Optional[Any] = "This is a simple input" snake_case__ : Tuple = ["This is a simple input 1", "This is a simple input 2"] snake_case__ : Any = ("This is a simple input", "This is a pair") snake_case__ : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(_SCREAMING_SNAKE_CASE , tokenizer_r.encode , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" ) # Simple input self.assertRaises(_SCREAMING_SNAKE_CASE , tokenizer_r.encode_plus , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" ) # Simple input self.assertRaises( _SCREAMING_SNAKE_CASE , tokenizer_r.batch_encode_plus , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" , ) # Pair input self.assertRaises(_SCREAMING_SNAKE_CASE , tokenizer_r.encode , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" ) # Pair input self.assertRaises(_SCREAMING_SNAKE_CASE , tokenizer_r.encode_plus , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" ) # Pair input self.assertRaises( _SCREAMING_SNAKE_CASE , tokenizer_r.batch_encode_plus , _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding="max_length" , ) def _lowercase ( self : Any ): snake_case__ : Any = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>" ) # Simple input snake_case__ : Dict = "This is a simple input" snake_case__ : Optional[Any] = ["This is a simple input looooooooong", "This is a simple input"] snake_case__ : Dict = ("This is a simple input", "This is a pair") snake_case__ : Dict = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] snake_case__ : int = tokenizer.pad_token_id snake_case__ : int = tokenizer(_SCREAMING_SNAKE_CASE , padding="max_length" , max_length=3_0 , return_tensors="np" ) snake_case__ : Tuple = tokenizer(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , truncate=_SCREAMING_SNAKE_CASE , return_tensors="np" ) snake_case__ : Dict = tokenizer(*_SCREAMING_SNAKE_CASE , padding="max_length" , max_length=6_0 , return_tensors="np" ) snake_case__ : int = tokenizer(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , truncate=_SCREAMING_SNAKE_CASE , return_tensors="np" ) # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s["input_ids"] ) self.assertTrue(0 in out_s["attention_mask"] ) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0] ) self.assertFalse(0 in out_sa["attention_mask"][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1] ) self.assertTrue(0 in out_sa["attention_mask"][1] ) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p["input_ids"] ) self.assertTrue(0 in out_p["attention_mask"] ) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0] ) self.assertFalse(0 in out_pa["attention_mask"][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1] ) self.assertTrue(0 in out_pa["attention_mask"][1] ) def _lowercase ( self : List[str] ): snake_case__ : List[str] = "$$$" snake_case__ : Union[str, Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=_SCREAMING_SNAKE_CASE , add_bos_token=_SCREAMING_SNAKE_CASE ) snake_case__ : Tuple = "This is a simple input" snake_case__ : Dict = ["This is a simple input 1", "This is a simple input 2"] snake_case__ : str = tokenizer.bos_token_id snake_case__ : Tuple = tokenizer(_SCREAMING_SNAKE_CASE ) snake_case__ : str = tokenizer(_SCREAMING_SNAKE_CASE ) self.assertEqual(out_s.input_ids[0] , _SCREAMING_SNAKE_CASE ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case__ : int = tokenizer.decode(out_s.input_ids ) snake_case__ : List[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , _SCREAMING_SNAKE_CASE ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def _lowercase ( self : int ): snake_case__ : Dict = CodeGenTokenizer.from_pretrained("Salesforce/codegen-350M-mono" ) snake_case__ : List[Any] = "\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#" snake_case__ : List[Any] = "\nif len_a > len_b: result = a\nelse: result = b" snake_case__ : Any = tokenizer.encode(_SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = ["^#", re.escape("<|endoftext|>" ), "^\'\'\'", "^\"\"\"", "\n\n\n"] snake_case__ : Optional[int] = tokenizer.decode(_SCREAMING_SNAKE_CASE , truncate_before_pattern=_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _lowercase ( self : Tuple ): pass
368
from datetime import datetime import matplotlib.pyplot as plt import torch def SCREAMING_SNAKE_CASE ( snake_case_ : int ): for param in module.parameters(): snake_case__ : Tuple = False def SCREAMING_SNAKE_CASE ( ): snake_case__ : Any = "cuda" if torch.cuda.is_available() else "cpu" if torch.backends.mps.is_available() and torch.backends.mps.is_built(): snake_case__ : List[Any] = "mps" if device == "mps": print( "WARNING: MPS currently doesn't seem to work, and messes up backpropagation without any visible torch" " errors. I recommend using CUDA on a colab notebook or CPU instead if you're facing inexplicable issues" " with generations." ) return device def SCREAMING_SNAKE_CASE ( snake_case_ : str ): snake_case__ : List[str] = plt.imshow(snake_case_ ) fig.axes.get_xaxis().set_visible(snake_case_ ) fig.axes.get_yaxis().set_visible(snake_case_ ) plt.show() def SCREAMING_SNAKE_CASE ( ): snake_case__ : str = datetime.now() snake_case__ : Optional[Any] = current_time.strftime("%H:%M:%S" ) return timestamp
286
0
"""simple docstring""" import math import random def __a ( __lowerCamelCase, __lowerCamelCase = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _a = 0.02 def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__UpperCamelCase ): # Forward propagation UpperCAmelCase_ : Tuple = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase_ : Optional[int] = (expected / 100) - layer_a # Error delta UpperCAmelCase_ : Optional[Any] = layer_1_error * sigmoid_function(__UpperCamelCase, __UpperCamelCase ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _a = int(input('Expected value: ')) _a = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
61
'''simple docstring''' def lowercase__ ( __UpperCamelCase = 4000000 )-> int: UpperCamelCase = [] UpperCamelCase ,UpperCamelCase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__UpperCamelCase ) UpperCamelCase ,UpperCamelCase = b, a + b return sum(__UpperCamelCase ) if __name__ == "__main__": print(f'{solution() = }')
321
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "SCUT-DLVCLab/lilt-roberta-en-base": ( "https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json" ), } class UpperCAmelCase_ ( A_ ): lowercase__ = '''lilt''' def __init__( self : List[str] , snake_case_ : Any=30_522 , snake_case_ : Optional[Any]=768 , snake_case_ : Union[str, Any]=12 , snake_case_ : Union[str, Any]=12 , snake_case_ : Any=3_072 , snake_case_ : List[str]="gelu" , snake_case_ : Union[str, Any]=0.1 , snake_case_ : List[Any]=0.1 , snake_case_ : Any=512 , snake_case_ : Optional[Any]=2 , snake_case_ : List[str]=0.02 , snake_case_ : Optional[Any]=1e-12 , snake_case_ : List[Any]=0 , snake_case_ : Any="absolute" , snake_case_ : str=None , snake_case_ : int=4 , snake_case_ : int=1_024 , **snake_case_ : Tuple , ) -> List[Any]: '''simple docstring''' super().__init__(pad_token_id=snake_case_ , **snake_case_ ) A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_act A__ = intermediate_size A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = initializer_range A__ = layer_norm_eps A__ = position_embedding_type A__ = classifier_dropout A__ = channel_shrink_ratio A__ = max_ad_position_embeddings
230
"""simple docstring""" from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class UpperCAmelCase_ ( nn.Module ): def __init__( self : Optional[int] , snake_case_ : int = 16 , snake_case_ : int = 88 , snake_case_ : Optional[int] = None , snake_case_ : int = 1 , snake_case_ : float = 0.0 , snake_case_ : int = 32 , snake_case_ : Optional[int] = None , snake_case_ : bool = False , snake_case_ : Optional[int] = None , snake_case_ : Optional[int] = None , snake_case_ : str = "geglu" , snake_case_ : Optional[int] = None , ) -> str: '''simple docstring''' super().__init__() A__ = nn.ModuleList( [ TransformeraDModel( num_attention_heads=snake_case_ , attention_head_dim=snake_case_ , in_channels=snake_case_ , num_layers=snake_case_ , dropout=snake_case_ , norm_num_groups=snake_case_ , cross_attention_dim=snake_case_ , attention_bias=snake_case_ , sample_size=snake_case_ , num_vector_embeds=snake_case_ , activation_fn=snake_case_ , num_embeds_ada_norm=snake_case_ , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference A__ = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` A__ = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` A__ = [1, 0] def __magic_name__ ( self : Dict , snake_case_ : List[Any] , snake_case_ : Tuple , snake_case_ : Any=None , snake_case_ : int=None , snake_case_ : Union[str, Any]=None , snake_case_ : bool = True , ) -> Union[str, Any]: '''simple docstring''' A__ = hidden_states A__ = [] A__ = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens A__ = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] A__ = self.transformer_index_for_condition[i] A__ = self.transformers[transformer_index]( snake_case_ , encoder_hidden_states=snake_case_ , timestep=snake_case_ , cross_attention_kwargs=snake_case_ , return_dict=snake_case_ , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] A__ = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) A__ = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=snake_case_ )
230
1
import os from datetime import datetime as dt from github import Github UpperCamelCase = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''enhancement''', '''new pipeline/model''', '''new scheduler''', '''wip''', ] def lowercase_ ( ): lowercase__ : Optional[Any] = Github(os.environ["GITHUB_TOKEN"]) lowercase__ : Dict = g.get_repo("huggingface/diffusers") lowercase__ : int = repo.get_issues(state="open") for issue in open_issues: lowercase__ : str = sorted(issue.get_comments() , key=lambda _lowerCamelCase: i.created_at , reverse=_lowerCamelCase) lowercase__ : int = comments[0] if len(_lowerCamelCase) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="closed") elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="open") issue.remove_from_labels("stale") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") issue.add_to_labels("stale") if __name__ == "__main__": main()
87
import argparse import torch from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert from transformers.utils import logging logging.set_verbosity_info() def lowercase_ ( _lowerCamelCase : str , _lowerCamelCase : List[Any] , _lowerCamelCase : Dict): # Initialise PyTorch model lowercase__ : List[str] = BertConfig.from_json_file(_lowerCamelCase) print(f'''Building PyTorch model from configuration: {config}''') lowercase__ : Optional[Any] = BertForPreTraining(_lowerCamelCase) # Load weights from tf checkpoint load_tf_weights_in_bert(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''') torch.save(model.state_dict() , _lowerCamelCase) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--bert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
87
1
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _lowerCAmelCase = parse(importlib.metadata.version('''torch''')) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) lowerCAmelCase__ : int = STR_OPERATION_TO_FUNC[operation] if isinstance(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : List[Any] = parse(importlib.metadata.version(UpperCamelCase ) ) return operation(UpperCamelCase , parse(UpperCamelCase ) ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return compare_versions(UpperCamelCase , UpperCamelCase , UpperCamelCase )
184
'''simple docstring''' from PIL import Image def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ : Dict = image.size lowerCAmelCase__ : int = 0 lowerCAmelCase__ : int = image.load() for i in range(UpperCamelCase ): for j in range(UpperCamelCase ): lowerCAmelCase__ : int = pixels[j, i] mean += pixel mean //= width * height for j in range(UpperCamelCase ): for i in range(UpperCamelCase ): lowerCAmelCase__ : Dict = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": _lowerCAmelCase = mean_threshold(Image.open('''path_to_image''').convert('''L''')) image.save('''output_image_path''')
184
1
from __future__ import annotations from typing import Any def _a ( lowerCamelCase: list[Any] ) -> None: '''simple docstring''' create_state_space_tree(lowerCamelCase , [] , 0 ) def _a ( lowerCamelCase: list[Any] , lowerCamelCase: list[Any] , lowerCamelCase: int ) -> None: '''simple docstring''' if index == len(lowerCamelCase ): print(lowerCamelCase ) return create_state_space_tree(lowerCamelCase , lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowerCamelCase , lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": snake_case__ : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['A', 'B', 'C']) generate_all_subsequences(seq)
117
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass snake_case__ : Union[str, Any] = (3, 9, -11, 0, 7, 5, 1, -1) snake_case__ : int = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class A_ : lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 class A_ : def __init__(self :Dict , _UpperCamelCase :Iterable[int] )-> None: __A = None for i in sorted(_UpperCamelCase , reverse=_UpperCamelCase ): __A = Node(_UpperCamelCase , self.head ) def __iter__(self :List[str] )-> Iterator[int]: __A = self.head while node: yield node.data __A = node.next_node def __len__(self :Union[str, Any] )-> int: return sum(1 for _ in self ) def __str__(self :List[Any] )-> str: return " -> ".join([str(_UpperCamelCase ) for node in self] ) def _a ( lowerCamelCase: SortedLinkedList , lowerCamelCase: SortedLinkedList ) -> SortedLinkedList: '''simple docstring''' return SortedLinkedList(list(lowerCamelCase ) + list(lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() snake_case__ : Any = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
117
1
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _lowerCAmelCase : List[str] = logging.get_logger(__name__) class A_ : """simple docstring""" def __init__( self: List[Any] ,__lowerCAmelCase: str = None ,__lowerCAmelCase: uuid.UUID = None ,__lowerCAmelCase: Optional[Any]=None ,__lowerCAmelCase: Tuple=None ): '''simple docstring''' if not conversation_id: _lowerCamelCase : List[str] = uuid.uuida() if past_user_inputs is None: _lowerCamelCase : Dict = [] if generated_responses is None: _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : uuid.UUID = conversation_id _lowerCamelCase : List[str] = past_user_inputs _lowerCamelCase : List[str] = generated_responses _lowerCamelCase : Optional[str] = text def __eq__( self: Optional[int] ,__lowerCAmelCase: List[str] ): '''simple docstring''' if not isinstance(__lowerCAmelCase ,__lowerCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def _lowercase ( self: Dict ,__lowerCAmelCase: str ,__lowerCAmelCase: bool = False ): '''simple docstring''' if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) _lowerCamelCase : Tuple = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _lowerCamelCase : Optional[int] = text def _lowercase ( self: Tuple ): '''simple docstring''' if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _lowerCamelCase : str = None def _lowercase ( self: List[str] ,__lowerCAmelCase: str ): '''simple docstring''' self.generated_responses.append(__lowerCAmelCase ) def _lowercase ( self: Any ): '''simple docstring''' for user_input, generated_response in zip(self.past_user_inputs ,self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: List[str] ): '''simple docstring''' _lowerCamelCase : Dict = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _lowerCamelCase : Dict = "user" if is_user else "bot" output += F"""{name} >> {text} \n""" return output @add_end_docstrings( _a , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class A_ ( _a ): """simple docstring""" def __init__( self: Dict ,*__lowerCAmelCase: Dict ,**__lowerCAmelCase: Any ): '''simple docstring''' super().__init__(*__lowerCAmelCase ,**__lowerCAmelCase ) if self.tokenizer.pad_token_id is None: _lowerCamelCase : Tuple = self.tokenizer.eos_token def _lowercase ( self: Optional[int] ,__lowerCAmelCase: List[str]=None ,__lowerCAmelCase: List[Any]=None ,__lowerCAmelCase: Tuple=None ,**__lowerCAmelCase: Tuple ): '''simple docstring''' _lowerCamelCase : Optional[Any] = {} _lowerCamelCase : Any = {} _lowerCamelCase : Tuple = {} if min_length_for_response is not None: _lowerCamelCase : Union[str, Any] = min_length_for_response if minimum_tokens is not None: _lowerCamelCase : Dict = minimum_tokens if "max_length" in generate_kwargs: _lowerCamelCase : Optional[Any] = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _lowerCamelCase : Union[str, Any] = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(__lowerCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self: Union[str, Any] ,__lowerCAmelCase: Union[Conversation, List[Conversation]] ,__lowerCAmelCase: Optional[int]=0 ,**__lowerCAmelCase: Any ): '''simple docstring''' _lowerCamelCase : int = super().__call__(__lowerCAmelCase ,num_workers=__lowerCAmelCase ,**__lowerCAmelCase ) if isinstance(__lowerCAmelCase ,__lowerCAmelCase ) and len(__lowerCAmelCase ) == 1: return outputs[0] return outputs def _lowercase ( self: Any ,__lowerCAmelCase: Conversation ,__lowerCAmelCase: List[str]=32 ): '''simple docstring''' if not isinstance(__lowerCAmelCase ,__lowerCAmelCase ): raise ValueError("ConversationalPipeline, expects Conversation as inputs" ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ "Add user inputs with the conversation's `add_user_input` method" ) if hasattr(self.tokenizer ,"_build_conversation_input_ids" ): _lowerCamelCase : Tuple = self.tokenizer._build_conversation_input_ids(__lowerCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _lowerCamelCase : Optional[int] = self._legacy_parse_and_tokenize(__lowerCAmelCase ) if self.framework == "pt": _lowerCamelCase : int = torch.LongTensor([input_ids] ) elif self.framework == "tf": _lowerCamelCase : Tuple = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def _lowercase ( self: List[Any] ,__lowerCAmelCase: Tuple ,__lowerCAmelCase: Tuple=10 ,**__lowerCAmelCase: Dict ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = generate_kwargs.get("max_length" ,self.model.config.max_length ) _lowerCamelCase : Any = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _lowerCamelCase : Union[str, Any] = max_length - minimum_tokens _lowerCamelCase : Optional[Any] = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _lowerCamelCase : Dict = model_inputs["attention_mask"][:, -trim:] _lowerCamelCase : str = model_inputs.pop("conversation" ) _lowerCamelCase : Tuple = max_length _lowerCamelCase : Union[str, Any] = self.model.generate(**__lowerCAmelCase ,**__lowerCAmelCase ) if self.model.config.is_encoder_decoder: _lowerCamelCase : Dict = 1 else: _lowerCamelCase : Any = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def _lowercase ( self: Optional[Any] ,__lowerCAmelCase: int ,__lowerCAmelCase: str=True ): '''simple docstring''' _lowerCamelCase : Tuple = model_outputs["output_ids"] _lowerCamelCase : Optional[Any] = self.tokenizer.decode( output_ids[0] ,skip_special_tokens=__lowerCAmelCase ,clean_up_tokenization_spaces=__lowerCAmelCase ,) _lowerCamelCase : List[str] = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(__lowerCAmelCase ) return conversation def _lowercase ( self: List[Any] ,__lowerCAmelCase: Conversation ): '''simple docstring''' _lowerCamelCase : Tuple = self.tokenizer.eos_token_id _lowerCamelCase : int = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(__lowerCAmelCase ,add_special_tokens=__lowerCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(__lowerCAmelCase ,add_special_tokens=__lowerCAmelCase ) ) if len(__lowerCAmelCase ) > self.tokenizer.model_max_length: _lowerCamelCase : List[Any] = input_ids[-self.tokenizer.model_max_length :] return input_ids
370
"""simple docstring""" _lowerCAmelCase : Tuple = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Tuple: '''simple docstring''' _lowerCamelCase : Any = [False] * len(_lowerCamelCase ) _lowerCamelCase : Union[str, Any] = [s] _lowerCamelCase : str = True while queue: _lowerCamelCase : Optional[int] = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_lowerCamelCase ) _lowerCamelCase : Any = True _lowerCamelCase : Any = u return visited[t] def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase : List[str] = [-1] * (len(_lowerCamelCase )) _lowerCamelCase : Union[str, Any] = 0 _lowerCamelCase : Union[str, Any] = [] _lowerCamelCase : List[str] = [i[:] for i in graph] # Record original cut, copy. while bfs(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): _lowerCamelCase : Any = float("Inf" ) _lowerCamelCase : Dict = sink while s != source: # Find the minimum value in select path _lowerCamelCase : Union[str, Any] = min(_lowerCamelCase , graph[parent[s]][s] ) _lowerCamelCase : Union[str, Any] = parent[s] max_flow += path_flow _lowerCamelCase : Optional[Any] = sink while v != source: _lowerCamelCase : Union[str, Any] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow _lowerCamelCase : List[str] = parent[v] for i in range(len(_lowerCamelCase ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
340
0
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE__ : str = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
48
def A ( _SCREAMING_SNAKE_CASE ) -> list: if n_term == "": return [] lowerCamelCase : list = [] for temp in range(int(_SCREAMING_SNAKE_CASE ) ): series.append(f'''1/{temp + 1}''' if series else "1" ) return series if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Dict = input('Enter the last number (nth term) of the Harmonic Series') print('Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n') print(harmonic_series(nth_term))
48
1
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 UpperCamelCase_ = { "return_dict": False, "output_hidden_states": True, "output_attentions": True, "torchscript": True, "torch_dtype": "float16", "use_bfloat16": True, "tf_legacy_loss": True, "pruned_heads": {"a": 1}, "tie_word_embeddings": False, "is_decoder": True, "cross_attention_hidden_size": 1_2_8, "add_cross_attention": True, "tie_encoder_decoder": True, "max_length": 5_0, "min_length": 3, "do_sample": True, "early_stopping": True, "num_beams": 3, "num_beam_groups": 3, "diversity_penalty": 0.5, "temperature": 2.0, "top_k": 1_0, "top_p": 0.7, "typical_p": 0.2, "repetition_penalty": 0.8, "length_penalty": 0.8, "no_repeat_ngram_size": 5, "encoder_no_repeat_ngram_size": 5, "bad_words_ids": [1, 2, 3], "num_return_sequences": 3, "chunk_size_feed_forward": 5, "output_scores": True, "return_dict_in_generate": True, "forced_bos_token_id": 2, "forced_eos_token_id": 3, "remove_invalid_values": True, "architectures": ["BertModel"], "finetuning_task": "translation", "id2label": {0: "label"}, "label2id": {"label": "0"}, "tokenizer_class": "BertTokenizerFast", "prefix": "prefix", "bos_token_id": 6, "pad_token_id": 7, "eos_token_id": 8, "sep_token_id": 9, "decoder_start_token_id": 1_0, "exponential_decay_length_penalty": (5, 1.01), "suppress_tokens": [0, 1], "begin_suppress_tokens": 2, "task_specific_params": {"translation": "some_params"}, "problem_type": "regression", } @is_staging_test class _a ( unittest.TestCase ): '''simple docstring''' @classmethod def UpperCamelCase_ ( cls ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def UpperCamelCase_ ( cls ): '''simple docstring''' try: delete_repo(token=cls._token, repo_id='test-config' ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id='valid_org/test-config-org' ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id='test-dynamic-config' ) except HTTPError: pass def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = BertConfig( vocab_size=99, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, intermediate_size=37 ) config.push_to_hub('test-config', use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Tuple = BertConfig.from_pretrained(F"{USER}/test-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_, getattr(lowercase_, lowercase_ ) ) # Reset repo delete_repo(token=self._token, repo_id='test-config' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowercase_, repo_id='test-config', push_to_hub=lowercase_, use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Optional[int] = BertConfig.from_pretrained(F"{USER}/test-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_, getattr(lowercase_, lowercase_ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = BertConfig( vocab_size=99, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, intermediate_size=37 ) config.push_to_hub('valid_org/test-config-org', use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Dict = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_, getattr(lowercase_, lowercase_ ) ) # Reset repo delete_repo(token=self._token, repo_id='valid_org/test-config-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowercase_, repo_id='valid_org/test-config-org', push_to_hub=lowercase_, use_auth_token=self._token ) SCREAMING_SNAKE_CASE : Union[str, Any] = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_, getattr(lowercase_, lowercase_ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' CustomConfig.register_for_auto_class() SCREAMING_SNAKE_CASE : int = CustomConfig(attribute=42 ) config.push_to_hub('test-dynamic-config', use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map, {'AutoConfig': 'custom_configuration.CustomConfig'} ) SCREAMING_SNAKE_CASE : str = AutoConfig.from_pretrained(F"{USER}/test-dynamic-config", trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__, 'CustomConfig' ) self.assertEqual(new_config.attribute, 42 ) class _a ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated SCREAMING_SNAKE_CASE : int = c.n_embd + 1 # int SCREAMING_SNAKE_CASE : int = c.resid_pdrop + 1.0 # float SCREAMING_SNAKE_CASE : List[str] = not c.scale_attn_weights # bool SCREAMING_SNAKE_CASE : List[str] = c.summary_type + 'foo' # str c.update_from_string( F"n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}" ) self.assertEqual(lowercase_, c.n_embd, 'mismatch for key: n_embd' ) self.assertEqual(lowercase_, c.resid_pdrop, 'mismatch for key: resid_pdrop' ) self.assertEqual(lowercase_, c.scale_attn_weights, 'mismatch for key: scale_attn_weights' ) self.assertEqual(lowercase_, c.summary_type, 'mismatch for key: summary_type' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = PretrainedConfig() SCREAMING_SNAKE_CASE : Dict = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowercase_, ['is_encoder_decoder', '_name_or_path', '_commit_hash', 'transformers_version'] ) SCREAMING_SNAKE_CASE : str = [key for key, value in config_common_kwargs.items() if value == getattr(lowercase_, lowercase_ )] if len(lowercase_ ) > 0: raise ValueError( 'The following keys are set with the default values in' ' `test_configuration_common.config_common_kwargs` pick another value for them:' F" {', '.join(lowercase_ )}." ) def UpperCamelCase_ ( self ): '''simple docstring''' with self.assertRaises(lowercase_ ): # config is in subfolder, the following should not work without specifying the subfolder SCREAMING_SNAKE_CASE : Any = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder' ) SCREAMING_SNAKE_CASE : Optional[int] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder', subfolder='bert' ) self.assertIsNotNone(lowercase_ ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = mock.Mock() SCREAMING_SNAKE_CASE : List[Any] = 500 SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Any = HTTPError SCREAMING_SNAKE_CASE : Dict = {} # Download this model to make sure it's in the cache. SCREAMING_SNAKE_CASE : Tuple = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request', return_value=lowercase_ ) as mock_head: SCREAMING_SNAKE_CASE : Dict = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = BertConfig.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = AutoConfig.from_pretrained('bert-base-cased' ) SCREAMING_SNAKE_CASE : int = ['config.4.0.0.json'] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowercase_ ) SCREAMING_SNAKE_CASE : Dict = 2 json.dump(configuration.to_dict(), open(os.path.join(lowercase_, 'config.4.0.0.json' ), 'w' ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 SCREAMING_SNAKE_CASE : str = AutoConfig.from_pretrained(lowercase_ ) self.assertEqual(new_configuration.hidden_size, 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 SCREAMING_SNAKE_CASE : Dict = ['config.42.0.0.json'] SCREAMING_SNAKE_CASE : Optional[int] = 768 configuration.save_pretrained(lowercase_ ) shutil.move(os.path.join(lowercase_, 'config.4.0.0.json' ), os.path.join(lowercase_, 'config.42.0.0.json' ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertEqual(new_configuration.hidden_size, 768 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = 'hf-internal-testing/test-two-configs' import transformers as new_transformers SCREAMING_SNAKE_CASE : List[Any] = 'v4.0.0' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = new_transformers.models.auto.AutoConfig.from_pretrained( lowercase_, return_unused_kwargs=lowercase_ ) self.assertEqual(new_configuration.hidden_size, 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowercase_, {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers SCREAMING_SNAKE_CASE : Dict = 'v3.0.0' SCREAMING_SNAKE_CASE : List[str] = old_transformers.models.auto.AutoConfig.from_pretrained(lowercase_ ) self.assertEqual(old_configuration.hidden_size, 768 )
364
'''simple docstring''' import math def lowercase__( __UpperCamelCase: float ,__UpperCamelCase: float ): """simple docstring""" return math.pow(__UpperCamelCase ,2 ) - a def lowercase__( __UpperCamelCase: float ): """simple docstring""" return 2 * x def lowercase__( __UpperCamelCase: float ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = 2.0 while start <= a: SCREAMING_SNAKE_CASE : Dict = math.pow(__UpperCamelCase ,2 ) return start def lowercase__( __UpperCamelCase: float ,__UpperCamelCase: int = 99_99 ,__UpperCamelCase: float = 0.0_0_0_0_0_0_0_0_0_0_0_0_0_1 ): """simple docstring""" if a < 0: raise ValueError('math domain error' ) SCREAMING_SNAKE_CASE : Union[str, Any] = get_initial_point(__UpperCamelCase ) for _ in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE : Optional[int] = value SCREAMING_SNAKE_CASE : Dict = value - fx(__UpperCamelCase ,__UpperCamelCase ) / fx_derivative(__UpperCamelCase ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
246
0
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Dict = FileLock(str(tmpdir / "foo.lock" ) ) _lowerCAmelCase : Union[str, Any] = FileLock(str(tmpdir / "foo.lock" ) ) _lowerCAmelCase : Optional[int] = 0.01 with locka.acquire(): with pytest.raises(_lowerCamelCase ): _lowerCAmelCase : Optional[Any] = time.time() locka.acquire(_lowerCamelCase ) assert time.time() - _start > timeout def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Optional[int] = "a" * 1_000 + ".lock" _lowerCAmelCase : Union[str, Any] = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(".lock" ) assert not locka._lock_file.endswith(_lowerCamelCase ) assert len(os.path.basename(locka._lock_file ) ) <= 255 _lowerCAmelCase : str = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(_lowerCamelCase ): locka.acquire(0 )
36
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def A ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Union[str, Any]=False ) -> str: '''simple docstring''' try: _UpperCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _UpperCAmelCase = default else: # KEY is set, convert it to True or False. try: _UpperCAmelCase = strtobool(_UpperCAmelCase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F"If set, {key} must be yes or no." ) return _value UpperCAmelCase__ = parse_flag_from_env("RUN_SLOW", default=False) def A ( _UpperCAmelCase : List[str] ) -> List[str]: '''simple docstring''' return unittest.skip('Test was skipped' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Dict ) -> str: '''simple docstring''' return unittest.skipUnless(_run_slow_tests , 'test is slow' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> str: '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , 'test requires only a CPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Dict ) -> Dict: '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , 'test requires a GPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless(is_xpu_available() , 'test requires a XPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[int] ) -> List[str]: '''simple docstring''' return unittest.skipUnless(is_mps_available() , 'test requires a `mps` backend support in `torch`' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , 'test requires the Hugging Face suite' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : str ) -> str: '''simple docstring''' return unittest.skipUnless(is_bnb_available() , 'test requires the bitsandbytes library' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless(is_tpu_available() , 'test requires TPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[Any] ) -> str: '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , 'test requires a GPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Tuple ) -> int: '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , 'test requires a XPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , 'test requires multiple GPUs' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Tuple ) -> Any: '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , 'test requires multiple XPUs' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , 'test requires safetensors' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[Any] ) -> Dict: '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , 'test requires DeepSpeed' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[int] ) -> str: '''simple docstring''' return unittest.skipUnless(is_torch_version('>=' , '1.12.0' ) , 'test requires torch version >= 1.12.0' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any=None , _UpperCAmelCase : List[Any]=None ) -> Dict: '''simple docstring''' if test_case is None: return partial(_UpperCAmelCase , version=_UpperCAmelCase ) return unittest.skipUnless(is_torch_version('>=' , _UpperCAmelCase ) , F"test requires torch version >= {version}" )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[str] ) -> int: '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , 'test requires Tensorboard' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' return unittest.skipUnless(is_wandb_available() , 'test requires wandb' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[str] ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , 'test requires comet_ml' )(_UpperCAmelCase ) UpperCAmelCase__ = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def A ( _UpperCAmelCase : List[str] ) -> Any: '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , 'test requires at least one tracker to be available and for `comet_ml` to not be installed' , )(_UpperCAmelCase ) class __lowerCAmelCase ( unittest.TestCase ): UpperCamelCase = True @classmethod def _lowerCamelCase ( cls : List[Any]) -> Tuple: """simple docstring""" _UpperCAmelCase = tempfile.mkdtemp() @classmethod def _lowerCamelCase ( cls : Union[str, Any]) -> str: """simple docstring""" if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def _lowerCamelCase ( self : List[str]) -> List[Any]: """simple docstring""" if self.clear_on_setup: for path in Path(self.tmpdir).glob('**/*'): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A) class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Dict) -> Tuple: """simple docstring""" super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[int] , A : Union[mock.Mock, List[mock.Mock]]) -> Tuple: """simple docstring""" _UpperCAmelCase = mocks if isinstance(A , (tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def A ( _UpperCAmelCase : List[Any] ) -> int: '''simple docstring''' _UpperCAmelCase = AcceleratorState() _UpperCAmelCase = tensor[None].clone().to(state.device ) _UpperCAmelCase = gather(_UpperCAmelCase ).cpu() _UpperCAmelCase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , _UpperCAmelCase ): return False return True class __lowerCAmelCase : def __init__( self : Optional[Any] , A : Union[str, Any] , A : Optional[int] , A : str) -> Optional[int]: """simple docstring""" _UpperCAmelCase = returncode _UpperCAmelCase = stdout _UpperCAmelCase = stderr async def A ( _UpperCAmelCase : str , _UpperCAmelCase : Optional[int] ) -> Optional[Any]: '''simple docstring''' while True: _UpperCAmelCase = await stream.readline() if line: callback(_UpperCAmelCase ) else: break async def A ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : str=None , _UpperCAmelCase : str=None , _UpperCAmelCase : Dict=False , _UpperCAmelCase : Union[str, Any]=False ) -> _RunOutput: '''simple docstring''' if echo: print('\nRunning: ' , ' '.join(_UpperCAmelCase ) ) _UpperCAmelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_UpperCAmelCase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_UpperCAmelCase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _UpperCAmelCase = [] _UpperCAmelCase = [] def tee(_UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : str="" ): _UpperCAmelCase = line.decode('utf-8' ).rstrip() sink.append(_UpperCAmelCase ) if not quiet: print(_UpperCAmelCase , _UpperCAmelCase , file=_UpperCAmelCase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stdout , label='stdout:' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stderr , label='stderr:' ) ) ), ] , timeout=_UpperCAmelCase , ) return _RunOutput(await p.wait() , _UpperCAmelCase , _UpperCAmelCase ) def A ( _UpperCAmelCase : str , _UpperCAmelCase : Dict=None , _UpperCAmelCase : str=None , _UpperCAmelCase : str=180 , _UpperCAmelCase : List[Any]=False , _UpperCAmelCase : List[Any]=True ) -> _RunOutput: '''simple docstring''' _UpperCAmelCase = asyncio.get_event_loop() _UpperCAmelCase = loop.run_until_complete( _stream_subprocess(_UpperCAmelCase , env=_UpperCAmelCase , stdin=_UpperCAmelCase , timeout=_UpperCAmelCase , quiet=_UpperCAmelCase , echo=_UpperCAmelCase ) ) _UpperCAmelCase = ' '.join(_UpperCAmelCase ) if result.returncode > 0: _UpperCAmelCase = '\n'.join(result.stderr ) raise RuntimeError( F"'{cmd_str}' failed with returncode {result.returncode}\n\n" F"The combined stderr from workers follows:\n{stderr}" ) return result class __lowerCAmelCase ( A ): pass def A ( _UpperCAmelCase : List[str] , _UpperCAmelCase : str=False ) -> Tuple: '''simple docstring''' try: _UpperCAmelCase = subprocess.check_output(_UpperCAmelCase , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(_UpperCAmelCase , 'decode' ): _UpperCAmelCase = output.decode('utf-8' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F"Command `{' '.join(_UpperCAmelCase )}` failed with the following error:\n\n{e.output.decode()}" ) from e
339
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : Tuple = { 'RWKV/rwkv-4-169m-pile': 'https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json', 'RWKV/rwkv-4-430m-pile': 'https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json', 'RWKV/rwkv-4-1b5-pile': 'https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json', 'RWKV/rwkv-4-3b-pile': 'https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json', 'RWKV/rwkv-4-7b-pile': 'https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json', 'RWKV/rwkv-4-14b-pile': 'https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json', 'RWKV/rwkv-raven-1b5': 'https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json', 'RWKV/rwkv-raven-3b': 'https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json', 'RWKV/rwkv-raven-7b': 'https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json', 'RWKV/rwkv-raven-14b': 'https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json', } class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): snake_case__ : Optional[int] = 'rwkv' snake_case__ : Tuple = {'max_position_embeddings': 'context_length'} def __init__( self : str , __lowerCamelCase : Any=50_277 , __lowerCamelCase : Optional[int]=1_024 , __lowerCamelCase : Dict=4_096 , __lowerCamelCase : List[str]=32 , __lowerCamelCase : List[str]=None , __lowerCamelCase : int=None , __lowerCamelCase : Dict=1E-5 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Tuple=6 , __lowerCamelCase : Dict=False , __lowerCamelCase : Tuple=True , **__lowerCamelCase : List[Any] , ): UpperCamelCase :int = vocab_size UpperCamelCase :Dict = context_length UpperCamelCase :str = hidden_size UpperCamelCase :str = num_hidden_layers UpperCamelCase :List[str] = attention_hidden_size if attention_hidden_size is not None else hidden_size UpperCamelCase :Tuple = intermediate_size if intermediate_size is not None else 4 * hidden_size UpperCamelCase :Any = layer_norm_epsilon UpperCamelCase :Dict = rescale_every UpperCamelCase :List[str] = use_cache UpperCamelCase :Tuple = bos_token_id UpperCamelCase :Any = eos_token_id super().__init__( tie_word_embeddings=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
368
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) UpperCAmelCase_ : Optional[int] = {'''vocab_file''': '''sentencepiece.bpe.model'''} UpperCAmelCase_ : str = { '''vocab_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model''' ), }, } UpperCAmelCase_ : List[Any] = { '''moussaKam/mbarthez''': 10_24, '''moussaKam/barthez''': 10_24, '''moussaKam/barthez-orangesum-title''': 10_24, } UpperCAmelCase_ : List[str] = '''▁''' class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Optional[int] = VOCAB_FILES_NAMES snake_case__ : int = PRETRAINED_VOCAB_FILES_MAP snake_case__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ : Dict = ["""input_ids""", """attention_mask"""] def __init__( self : int , __lowerCamelCase : Dict , __lowerCamelCase : List[str]="<s>" , __lowerCamelCase : str="</s>" , __lowerCamelCase : List[str]="</s>" , __lowerCamelCase : Optional[int]="<s>" , __lowerCamelCase : Optional[int]="<unk>" , __lowerCamelCase : Optional[Any]="<pad>" , __lowerCamelCase : str="<mask>" , __lowerCamelCase : Optional[Dict[str, Any]] = None , **__lowerCamelCase : Any , ): # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase :int = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else mask_token UpperCamelCase :int = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , sep_token=__lowerCamelCase , cls_token=__lowerCamelCase , pad_token=__lowerCamelCase , mask_token=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCamelCase , ) UpperCamelCase :Union[str, Any] = vocab_file UpperCamelCase :List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__lowerCamelCase ) ) UpperCamelCase :Optional[int] = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} UpperCamelCase :Tuple = len(self.sp_model ) - 1 UpperCamelCase :List[str] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def _A ( self : Optional[Any] , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase :Any = [self.cls_token_id] UpperCamelCase :Dict = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _A ( self : Dict , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) if token_ids_a is None: return [1] + ([0] * len(__lowerCamelCase )) + [1] return [1] + ([0] * len(__lowerCamelCase )) + [1, 1] + ([0] * len(__lowerCamelCase )) + [1] def _A ( self : int , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): UpperCamelCase :Any = [self.sep_token_id] UpperCamelCase :int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _A ( self : List[Any] ): return len(self.sp_model ) def _A ( self : Any ): UpperCamelCase :Optional[int] = {self.convert_ids_to_tokens(__lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _A ( self : int , __lowerCamelCase : str ): return self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase ) def _A ( self : Dict , __lowerCamelCase : Optional[int] ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCamelCase :List[Any] = self.sp_model.PieceToId(__lowerCamelCase ) return spm_id if spm_id else self.unk_token_id def _A ( self : Union[str, Any] , __lowerCamelCase : Union[str, Any] ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__lowerCamelCase ) def _A ( self : Optional[int] , __lowerCamelCase : Union[str, Any] ): UpperCamelCase :List[Any] = [] UpperCamelCase :str = """""" UpperCamelCase :Optional[int] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__lowerCamelCase ) + token UpperCamelCase :List[str] = True UpperCamelCase :Dict = [] else: current_sub_tokens.append(__lowerCamelCase ) UpperCamelCase :Optional[Any] = False out_string += self.sp_model.decode(__lowerCamelCase ) return out_string.strip() def __getstate__( self : str ): UpperCamelCase :Tuple = self.__dict__.copy() UpperCamelCase :str = None return state def __setstate__( self : Tuple , __lowerCamelCase : Optional[int] ): UpperCamelCase :Any = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCamelCase :Any = {} UpperCamelCase :str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A ( self : str , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ): if not os.path.isdir(__lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase :Union[str, Any] = os.path.join( __lowerCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCamelCase , """wb""" ) as fi: UpperCamelCase :List[str] = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase ) return (out_vocab_file,)
62
0
'''simple docstring''' from __future__ import annotations import pandas as pd def __lowercase ( __lowercase , __lowercase , __lowercase ) -> list[int]: '''simple docstring''' _A = [0] * no_of_processes _A = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(__lowercase ): _A = burst_time[i] _A = 0 _A = 0 _A = 9_9999_9999 _A = 0 _A = False # Process until all processes are completed while complete != no_of_processes: for j in range(__lowercase ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: _A = remaining_time[j] _A = j _A = True if not check: increment_time += 1 continue remaining_time[short] -= 1 _A = remaining_time[short] if minm == 0: _A = 9_9999_9999 if remaining_time[short] == 0: complete += 1 _A = False # Find finish time of current process _A = increment_time + 1 # Calculate waiting time _A = finish_time - arrival_time[short] _A = finar - burst_time[short] if waiting_time[short] < 0: _A = 0 # Increment time increment_time += 1 return waiting_time def __lowercase ( __lowercase , __lowercase , __lowercase ) -> list[int]: '''simple docstring''' _A = [0] * no_of_processes for i in range(__lowercase ): _A = burst_time[i] + waiting_time[i] return turn_around_time def __lowercase ( __lowercase , __lowercase , __lowercase ) -> None: '''simple docstring''' _A = 0 _A = 0 for i in range(__lowercase ): _A = total_waiting_time + waiting_time[i] _A = total_turn_around_time + turn_around_time[i] print(F'''Average waiting time = {total_waiting_time / no_of_processes:.5f}''' ) print("Average turn around time =" , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print('''Enter how many process you want to analyze''') lowerCamelCase_ = int(input()) lowerCamelCase_ = [0] * no_of_processes lowerCamelCase_ = [0] * no_of_processes lowerCamelCase_ = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print('''Enter the arrival time and burst time for process:--''' + str(i + 1)) lowerCamelCase_ , lowerCamelCase_ = map(int, input().split()) lowerCamelCase_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowerCamelCase_ = burst_time lowerCamelCase_ = no_of_processes lowerCamelCase_ = waiting_time lowerCamelCase_ = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) lowerCamelCase_ = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ '''Process''', '''BurstTime''', '''ArrivalTime''', '''WaitingTime''', '''TurnAroundTime''', ], ) # Printing the dataFrame pd.set_option('''display.max_rows''', fcfs.shape[0] + 1) print(fcfs)
79
'''simple docstring''' from __future__ import annotations from typing import Any class UpperCAmelCase_ : def __init__( self : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : float = 0 ) -> None: lowerCAmelCase , lowerCAmelCase = row, column lowerCAmelCase = [[default_value for c in range(UpperCAmelCase__ )] for r in range(UpperCAmelCase__ )] def __str__( self : List[str] ) -> str: lowerCAmelCase = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier lowerCAmelCase = 0 for row_vector in self.array: for obj in row_vector: lowerCAmelCase = max(UpperCAmelCase__ , len(str(UpperCAmelCase__ ) ) ) lowerCAmelCase = F'''%{max_element_length}s''' # Make string and return def single_line(UpperCAmelCase__ : list[float] ) -> str: nonlocal string_format_identifier lowerCAmelCase = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(UpperCAmelCase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ) -> str: return str(self ) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : tuple[int, int] ) -> bool: if not (isinstance(UpperCAmelCase__ , (list, tuple) ) and len(UpperCAmelCase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Any , UpperCAmelCase__ : tuple[int, int] ) -> Any: assert self.validate_indicies(UpperCAmelCase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Dict , UpperCAmelCase__ : tuple[int, int] , UpperCAmelCase__ : float ) -> None: assert self.validate_indicies(UpperCAmelCase__ ) lowerCAmelCase = value def __add__( self : Any , UpperCAmelCase__ : Matrix ) -> Matrix: assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert self.row == another.row and self.column == another.column # Add lowerCAmelCase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase = self[r, c] + another[r, c] return result def __neg__( self : int ) -> Matrix: lowerCAmelCase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase = -self[r, c] return result def __sub__( self : str , UpperCAmelCase__ : Matrix ) -> Matrix: return self + (-another) def __mul__( self : str , UpperCAmelCase__ : int | float | Matrix ) -> Matrix: if isinstance(UpperCAmelCase__ , (int, float) ): # Scalar multiplication lowerCAmelCase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase = self[r, c] * another return result elif isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): # Matrix multiplication assert self.column == another.row lowerCAmelCase = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: lowerCAmelCase = F'''Unsupported type given for another ({type(UpperCAmelCase__ )})''' raise TypeError(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[Any] ) -> Matrix: lowerCAmelCase = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase = self[r, c] return result def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : Matrix , UpperCAmelCase__ : Matrix ) -> Any: assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate lowerCAmelCase = v.transpose() lowerCAmelCase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def a_ ( ): # a^(-1) lowerCAmelCase = Matrix(3 , 3 , 0 ) for i in range(3 ): lowerCAmelCase = 1 print(f'''a^(-1) is {ainv}''' ) # u, v lowerCAmelCase = Matrix(3 , 1 , 0 ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 1, 2, -3 lowerCAmelCase = Matrix(3 , 1 , 0 ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase , lowerCamelCase )}''' ) def a_ ( ): import doctest doctest.testmod() testa()
4
0
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: """simple docstring""" snake_case_ : Union[str, Any] = 'laion/clap-htsat-unfused' snake_case_ : Union[str, Any] = tempfile.mkdtemp() def UpperCAmelCase_ ( self : Optional[int] , **_A : Optional[Any] ) -> Optional[Any]: """simple docstring""" return RobertaTokenizer.from_pretrained(self.checkpoint , **_A ) def UpperCAmelCase_ ( self : Tuple , **_A : List[str] ) -> int: """simple docstring""" return ClapFeatureExtractor.from_pretrained(self.checkpoint , **_A ) def UpperCAmelCase_ ( self : str ) -> Optional[Any]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" snake_case_ : Optional[Any] = self.get_tokenizer() snake_case_ : int = self.get_feature_extractor() snake_case_ : Optional[int] = ClapProcessor(tokenizer=_A , feature_extractor=_A ) processor.save_pretrained(self.tmpdirname ) snake_case_ : Optional[int] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def UpperCAmelCase_ ( self : Dict ) -> List[Any]: """simple docstring""" snake_case_ : Any = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) snake_case_ : Tuple = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) snake_case_ : Dict = self.get_feature_extractor(do_normalize=_A , padding_value=1.0 ) snake_case_ : List[str] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def UpperCAmelCase_ ( self : str ) -> str: """simple docstring""" snake_case_ : Optional[Any] = self.get_feature_extractor() snake_case_ : Optional[Any] = self.get_tokenizer() snake_case_ : Tuple = ClapProcessor(tokenizer=_A , feature_extractor=_A ) snake_case_ : Optional[Any] = floats_list((3, 1000) ) snake_case_ : Optional[int] = feature_extractor(_A , return_tensors='np' ) snake_case_ : Any = processor(audios=_A , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" snake_case_ : Any = self.get_feature_extractor() snake_case_ : List[Any] = self.get_tokenizer() snake_case_ : List[str] = ClapProcessor(tokenizer=_A , feature_extractor=_A ) snake_case_ : Union[str, Any] = 'This is a test string' snake_case_ : str = processor(text=_A ) snake_case_ : Any = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCAmelCase_ ( self : Any ) -> Union[str, Any]: """simple docstring""" snake_case_ : int = self.get_feature_extractor() snake_case_ : List[Any] = self.get_tokenizer() snake_case_ : Tuple = ClapProcessor(tokenizer=_A , feature_extractor=_A ) snake_case_ : int = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] snake_case_ : Union[str, Any] = processor.batch_decode(_A ) snake_case_ : int = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" snake_case_ : Optional[Any] = self.get_feature_extractor() snake_case_ : List[str] = self.get_tokenizer() snake_case_ : Dict = ClapProcessor(tokenizer=_A , feature_extractor=_A ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg='`processor` and `feature_extractor` model input names do not match' , )
88
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Optional[int] = None if token is not None: snake_case_ : List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} snake_case_ : Union[str, Any] = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100""" snake_case_ : Optional[int] = requests.get(__a , headers=__a ).json() snake_case_ : List[str] = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) snake_case_ : Dict = math.ceil((result['total_count'] - 1_00) / 1_00 ) for i in range(__a ): snake_case_ : Optional[Any] = requests.get(url + f"""&page={i + 2}""" , headers=__a ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Union[str, Any] = None if token is not None: snake_case_ : List[Any] = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} snake_case_ : Optional[Any] = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100""" snake_case_ : Union[str, Any] = requests.get(__a , headers=__a ).json() snake_case_ : Any = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) snake_case_ : str = math.ceil((result['total_count'] - 1_00) / 1_00 ) for i in range(__a ): snake_case_ : int = requests.get(url + f"""&page={i + 2}""" , headers=__a ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def SCREAMING_SNAKE_CASE__ ( __a , __a , __a , __a ): snake_case_ : Dict = None if token is not None: snake_case_ : List[Any] = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} snake_case_ : Optional[int] = requests.get(__a , headers=__a , allow_redirects=__a ) snake_case_ : str = result.headers['Location'] snake_case_ : List[str] = requests.get(__a , allow_redirects=__a ) snake_case_ : Optional[Any] = os.path.join(__a , f"""{artifact_name}.zip""" ) with open(__a , 'wb' ) as fp: fp.write(response.content ) def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Any = [] snake_case_ : Any = [] snake_case_ : Tuple = None with zipfile.ZipFile(__a ) as z: for filename in z.namelist(): if not os.path.isdir(__a ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(__a ) as f: for line in f: snake_case_ : Tuple = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs snake_case_ : Tuple = line[: line.index(': ' )] snake_case_ : Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed snake_case_ : Any = line[len('FAILED ' ) :] failed_tests.append(__a ) elif filename == "job_name.txt": snake_case_ : Union[str, Any] = line if len(__a ) != len(__a ): raise ValueError( f"""`errors` and `failed_tests` should have the same number of elements. Got {len(__a )} for `errors` """ f"""and {len(__a )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some""" ' problem.' ) snake_case_ : List[str] = None if job_name and job_links: snake_case_ : Union[str, Any] = job_links.get(__a , __a ) # A list with elements of the form (line of error, error, failed test) snake_case_ : Optional[Any] = [x + [y] + [job_link] for x, y in zip(__a , __a )] return result def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Any = [] snake_case_ : Any = [os.path.join(__a , __a ) for p in os.listdir(__a ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(__a , job_links=__a ) ) return errors def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Optional[int] = Counter() counter.update([x[1] for x in logs] ) snake_case_ : str = counter.most_common() snake_case_ : Tuple = {} for error, count in counts: if error_filter is None or error not in error_filter: snake_case_ : int = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} snake_case_ : int = dict(sorted(r.items() , key=lambda __a : item[1]["count"] , reverse=__a ) ) return r def SCREAMING_SNAKE_CASE__ ( __a ): snake_case_ : Tuple = test.split('::' )[0] if test.startswith('tests/models/' ): snake_case_ : List[str] = test.split('/' )[2] else: snake_case_ : Union[str, Any] = None return test def SCREAMING_SNAKE_CASE__ ( __a , __a=None ): snake_case_ : Optional[int] = [(x[0], x[1], get_model(x[2] )) for x in logs] snake_case_ : str = [x for x in logs if x[2] is not None] snake_case_ : int = {x[2] for x in logs} snake_case_ : Dict = {} for test in tests: snake_case_ : List[str] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) snake_case_ : Any = counter.most_common() snake_case_ : str = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} snake_case_ : Tuple = sum(error_counts.values() ) if n_errors > 0: snake_case_ : List[Any] = {'count': n_errors, 'errors': error_counts} snake_case_ : int = dict(sorted(r.items() , key=lambda __a : item[1]["count"] , reverse=__a ) ) return r def SCREAMING_SNAKE_CASE__ ( __a ): snake_case_ : Optional[Any] = '| no. | error | status |' snake_case_ : str = '|-:|:-|:-|' snake_case_ : Tuple = [header, sep] for error in reduced_by_error: snake_case_ : Dict = reduced_by_error[error]['count'] snake_case_ : List[str] = f"""| {count} | {error[:1_00]} | |""" lines.append(__a ) return "\n".join(__a ) def SCREAMING_SNAKE_CASE__ ( __a ): snake_case_ : Optional[Any] = '| model | no. of errors | major error | count |' snake_case_ : Union[str, Any] = '|-:|-:|-:|-:|' snake_case_ : Optional[int] = [header, sep] for model in reduced_by_model: snake_case_ : Any = reduced_by_model[model]['count'] snake_case_ ,snake_case_ : Dict = list(reduced_by_model[model]['errors'].items() )[0] snake_case_ : Any = f"""| {model} | {count} | {error[:60]} | {_count} |""" lines.append(__a ) return "\n".join(__a ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") _SCREAMING_SNAKE_CASE = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) _SCREAMING_SNAKE_CASE = get_job_links(args.workflow_run_id, token=args.token) _SCREAMING_SNAKE_CASE = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: _SCREAMING_SNAKE_CASE = k.find(""" / """) _SCREAMING_SNAKE_CASE = k[index + len(""" / """) :] _SCREAMING_SNAKE_CASE = v with open(os.path.join(args.output_dir, """job_links.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) _SCREAMING_SNAKE_CASE = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) _SCREAMING_SNAKE_CASE = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error _SCREAMING_SNAKE_CASE = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors _SCREAMING_SNAKE_CASE = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, """errors.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) _SCREAMING_SNAKE_CASE = reduce_by_error(errors) _SCREAMING_SNAKE_CASE = reduce_by_model(errors) _SCREAMING_SNAKE_CASE = make_github_table(reduced_by_error) _SCREAMING_SNAKE_CASE = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, """reduced_by_error.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa) with open(os.path.join(args.output_dir, """reduced_by_model.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa)
88
1
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _snake_case = logging.get_logger(__name__) @add_end_docstrings(a) class UpperCAmelCase_ ( a): def __init__( self, *__a, **__a): '''simple docstring''' super().__init__(*__a, **__a) requires_backends(self, "vision") self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == "tf" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING) def snake_case__ ( self, __a=None): '''simple docstring''' _lowerCAmelCase : Any = {} if top_k is not None: _lowerCAmelCase : Dict = top_k return {}, {}, postprocess_params def __call__( self, __a, **__a): '''simple docstring''' return super().__call__(__a, **__a) def snake_case__ ( self, __a): '''simple docstring''' _lowerCAmelCase : Dict = load_image(__a) _lowerCAmelCase : Tuple = self.image_processor(images=__a, return_tensors=self.framework) return model_inputs def snake_case__ ( self, __a): '''simple docstring''' _lowerCAmelCase : Dict = self.model(**__a) return model_outputs def snake_case__ ( self, __a, __a=5): '''simple docstring''' if top_k > self.model.config.num_labels: _lowerCAmelCase : Any = self.model.config.num_labels if self.framework == "pt": _lowerCAmelCase : int = model_outputs.logits.softmax(-1)[0] _lowerCAmelCase , _lowerCAmelCase : Any = probs.topk(__a) elif self.framework == "tf": _lowerCAmelCase : Optional[Any] = stable_softmax(model_outputs.logits, axis=-1)[0] _lowerCAmelCase : int = tf.math.top_k(__a, k=__a) _lowerCAmelCase , _lowerCAmelCase : int = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"Unsupported framework: {self.framework}") _lowerCAmelCase : Dict = scores.tolist() _lowerCAmelCase : List[Any] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__a, __a)]
36
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin UpperCAmelCase : Optional[int] = random.Random() def __lowerCamelCase ( lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[Any]=1.0 , lowerCamelCase__ : Union[str, Any]=None , lowerCamelCase__ : Dict=None ): '''simple docstring''' if rng is None: lowerCamelCase = global_rng lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self , A , A=7 , A=4_00 , A=20_00 , A=1 , A=0.0 , A=1_60_00 , A=True , A=True , ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = parent lowerCamelCase = batch_size lowerCamelCase = min_seq_length lowerCamelCase = max_seq_length lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase = feature_size lowerCamelCase = padding_value lowerCamelCase = sampling_rate lowerCamelCase = return_attention_mask lowerCamelCase = do_normalize def __A ( self ) -> Any: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __A ( self , A=False , A=False ) -> Any: '''simple docstring''' def _flatten(A ): return list(itertools.chain(*A ) ) if equal_length: lowerCamelCase = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCamelCase = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase = [np.asarray(A ) for x in speech_inputs] return speech_inputs class __lowercase ( a_ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Optional[int] = WavaVecaFeatureExtractor def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = WavaVecaFeatureExtractionTester(self ) def __A ( self , A ) -> Any: '''simple docstring''' self.assertTrue(np.all(np.mean(A , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A , axis=0 ) - 1 ) < 1e-3 ) ) def __A ( self ) -> Dict: '''simple docstring''' lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCamelCase = [np.asarray(A ) for speech_input in speech_inputs] # Test not batched input lowerCamelCase = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values lowerCamelCase = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(A , A , atol=1e-3 ) ) # Test batched lowerCamelCase = feat_extract(A , return_tensors="""np""" ).input_values lowerCamelCase = feat_extract(A , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(A , A ): self.assertTrue(np.allclose(A , A , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] lowerCamelCase = np.asarray(A ) lowerCamelCase = feat_extract(A , return_tensors="""np""" ).input_values lowerCamelCase = feat_extract(A , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(A , A ): self.assertTrue(np.allclose(A , A , atol=1e-3 ) ) def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCamelCase = ["""longest""", """max_length""", """do_not_pad"""] lowerCamelCase = [None, 16_00, None] for max_length, padding in zip(A , A ): lowerCamelCase = feat_extract(A , padding=A , max_length=A , return_tensors="""np""" ) lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self.assertTrue(input_values[0][8_00:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self.assertTrue(input_values[0][10_00:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase = range(8_00 , 14_00 , 2_00 ) lowerCamelCase = [floats_list((1, x) )[0] for x in lengths] lowerCamelCase = ["""longest""", """max_length""", """do_not_pad"""] lowerCamelCase = [None, 16_00, None] for max_length, padding in zip(A , A ): lowerCamelCase = feat_extract(A , max_length=A , padding=A ) lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def __A ( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCamelCase = feat_extract( A , truncation=A , max_length=10_00 , padding="""max_length""" , return_tensors="""np""" ) lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCamelCase = feat_extract( A , truncation=A , max_length=10_00 , padding="""longest""" , return_tensors="""np""" ) lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00) ) lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCamelCase = feat_extract( A , truncation=A , max_length=20_00 , padding="""longest""" , return_tensors="""np""" ) lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00) ) @require_torch def __A ( self ) -> Optional[int]: '''simple docstring''' import torch lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase = np.random.rand(1_00 ).astype(np.floataa ) lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCamelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def __A ( self ) -> str: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: lowerCamelCase = WavaVecaConfig.from_pretrained(A ) lowerCamelCase = WavaVecaFeatureExtractor.from_pretrained(A ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == """layer""" )
252
0
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __magic_name__ = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE_ ( __a , unittest.TestCase ): """simple docstring""" __lowercase : List[str] = XLNetTokenizer __lowercase : Union[str, Any] = XLNetTokenizerFast __lowercase : Tuple = True __lowercase : Union[str, Any] = True def snake_case_ ( self): super().setUp() # We have a SentencePiece fixture for testing __SCREAMING_SNAKE_CASE = XLNetTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = """<s>""" __SCREAMING_SNAKE_CASE = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__) , lowerCAmelCase__) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__) , lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , """<unk>""") self.assertEqual(vocab_keys[1] , """<s>""") self.assertEqual(vocab_keys[-1] , """<eod>""") self.assertEqual(len(lowerCAmelCase__) , 1_0_0_6) def snake_case_ ( self): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = XLNetTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("""This is a test""") self.assertListEqual(lowerCAmelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""]) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) __SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4]) __SCREAMING_SNAKE_CASE = tokenizer.convert_ids_to_tokens(lowerCAmelCase__) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = XLNetTokenizer(lowerCAmelCase__ , do_lower_case=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """""", """i""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] , ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""") , ["""▁he""", """ll""", """o"""]) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = XLNetTokenizer(lowerCAmelCase__ , do_lower_case=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] , ) @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = XLNetTokenizer.from_pretrained("""xlnet-base-cased""") __SCREAMING_SNAKE_CASE = tokenizer.encode("""sequence builders""" , add_special_tokens=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.encode("""multi-sequence build""" , add_special_tokens=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def snake_case_ ( self): # fmt: off __SCREAMING_SNAKE_CASE = {"""input_ids""": [[1_7, 2_1_4_4_2, 2_7_0, 1_7, 1_0, 1_4_6_4_5, 3_1_8, 3_4, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 7_7_5_2, 2_2_0_1_8, 2_3, 2_1, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 3_3_5_2, 1_4_4_3_1, 1_3, 5_5_0_0, 1_1, 1_1_7_6, 5_8_0, 1_3, 1_6_8_1_9, 4_7_9_7, 2_3, 1_7, 1_0, 1_7_1_3_5, 6_5_8, 1_9, 4_5_7, 7_9_3_2, 1_3, 1_8_4, 1_9, 3_1_5_4, 1_7_1_3_5, 6_4_6_8, 1_9, 1_4_0_4, 1_2_2_6_9, 1_9, 4_2_2_9, 5_3_5_6, 1_6_2_6_4, 4_6, 1_9, 1_7, 2_0_5_4_5, 1_0_3_9_5, 9, 9, 9, 1_1, 2_8, 6_4_2_1, 9_5_3_1, 2_0_7_2_9, 1_7, 1_0, 3_5_3, 1_7_0_2_2, 1_1, 2_1, 6_4_2_1, 9_5_3_1, 1_6_9_4_9, 1_7, 1_0, 1_1_5_0_9, 7_5_3, 1_1, 3_3, 9_5, 2_4_2_1, 7_3_8_5, 9_5_6, 1_4_4_3_1, 2_6_2_6, 2_5, 8_4_2, 7_3_8_5, 4_8_3_6, 2_1, 1_4_2_9, 2_2_7_2, 9_8_5_5, 3_1_2_0, 1_6_1, 2_4_7_3_8, 1_9, 1_3_2_0_3, 6_5_8, 2_1_8, 7_8_7, 2_1, 4_3_0, 1_8_4_8_2, 8_4_7, 2_6_3_7, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2_2, 2_2_1_7_8, 2_7, 1_0_6_4, 2_2, 9_5_6, 1_3, 1_1_1_0_1, 1_4_2_9, 5_8_5_4, 2_4_3_1_3, 1_8_9_5_3, 4_0, 4_2_2, 2_4_3_6_6, 6_8, 1_7_5_8, 3_7, 1_0_4_8_3, 1_4_2_5_7, 3_1, 2_0_7, 2_6_3, 2_1, 2_0_3, 3_7_7_3, 2_5, 7_1, 9_7_3_5, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2, 2_0_4_9, 3_4_4_2, 1_7, 1_3_8_9_4, 3_3_8_0, 2_3, 9_5, 1_8, 1_7_6_3_4, 2_2_8_8, 9, 4, 3]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="""xlnet-base-cased""" , revision="""c841166438c31ec7ca9a106dee7bb312b73ae511""" , )
355
"""simple docstring""" import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=1024 ): __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = [], [] __SCREAMING_SNAKE_CASE = list(zip(UpperCamelCase_ , UpperCamelCase_ ) ) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = sorted_examples[0] def is_too_big(UpperCamelCase_ ): return tok(UpperCamelCase_ , return_tensors="""pt""" ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): __SCREAMING_SNAKE_CASE = new_src + """ """ + src __SCREAMING_SNAKE_CASE = new_tgt + """ """ + tgt if is_too_big(UpperCamelCase_ ) or is_too_big(UpperCamelCase_ ): # cant fit, finalize example finished_src.append(UpperCamelCase_ ) finished_tgt.append(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = src, tgt else: # can fit, keep adding __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(UpperCamelCase_ ) finished_tgt.append(UpperCamelCase_ ) return finished_src, finished_tgt def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = Path(UpperCamelCase_ ) save_path.mkdir(exist_ok=UpperCamelCase_ ) for split in ["train"]: __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = data_dir / f"{split}.source", data_dir / f"{split}.target" __SCREAMING_SNAKE_CASE = [x.rstrip() for x in Path(UpperCamelCase_ ).open().readlines()] __SCREAMING_SNAKE_CASE = [x.rstrip() for x in Path(UpperCamelCase_ ).open().readlines()] __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = pack_examples(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) print(f"packed {split} split from {len(UpperCamelCase_ )} examples -> {len(UpperCamelCase_ )}." ) Path(save_path / f"{split}.source" ).open("""w""" ).write("""\n""".join(UpperCamelCase_ ) ) Path(save_path / f"{split}.target" ).open("""w""" ).write("""\n""".join(UpperCamelCase_ ) ) for split in ["val", "test"]: __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = data_dir / f"{split}.source", data_dir / f"{split}.target" shutil.copyfile(UpperCamelCase_ , save_path / f"{split}.source" ) shutil.copyfile(UpperCamelCase_ , save_path / f"{split}.target" ) def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument("""--tok_name""" , type=UpperCamelCase_ , help="""like facebook/bart-large-cnn,t5-base, etc.""" ) parser.add_argument("""--max_seq_len""" , type=UpperCamelCase_ , default=128 ) parser.add_argument("""--data_dir""" , type=UpperCamelCase_ ) parser.add_argument("""--save_path""" , type=UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = parser.parse_args() __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(UpperCamelCase_ , Path(args.data_dir ) , args.max_seq_len , args.save_path ) if __name__ == "__main__": packer_cli()
255
0
'''simple docstring''' from __future__ import annotations __SCREAMING_SNAKE_CASE : Optional[int] = list[list[int]] # assigning initial values to the grid __SCREAMING_SNAKE_CASE : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution __SCREAMING_SNAKE_CASE : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def UpperCamelCase_ ( _UpperCAmelCase : Matrix , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(_UpperCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : int = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): _UpperCAmelCase : Dict = digit if sudoku(_UpperCAmelCase ) is not None: return grid _UpperCAmelCase : Tuple = 0 return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(_UpperCAmelCase , end=" " ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("""\nExample grid:\n""" + """=""" * 20) print_solution(example_grid) print("""\nExample grid solution:""") __SCREAMING_SNAKE_CASE : Tuple = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("""Cannot find a solution.""")
31
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
0
def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ = 1_000 ): __lowerCamelCase : Optional[int] = 2**power __lowerCamelCase : Any = str(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Tuple = list(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Dict = 0 for i in list_num: sum_of_num += int(SCREAMING_SNAKE_CASE__ ) return sum_of_num if __name__ == "__main__": lowercase_ = int(input('Enter the power of 2: ').strip()) print('2 ^ ', power, ' = ', 2**power) lowercase_ = solution(power) print('Sum of the digits is: ', result)
194
def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : str = 0 __lowerCamelCase : Tuple = len(SCREAMING_SNAKE_CASE__ ) for i in range(n - 1 ): for j in range(i + 1 , SCREAMING_SNAKE_CASE__ ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): if len(SCREAMING_SNAKE_CASE__ ) <= 1: return arr, 0 __lowerCamelCase : Optional[int] = len(SCREAMING_SNAKE_CASE__ ) // 2 __lowerCamelCase : Union[str, Any] = arr[0:mid] __lowerCamelCase : List[Any] = arr[mid:] __lowerCamelCase , __lowerCamelCase : Any = count_inversions_recursive(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase , __lowerCamelCase : List[str] = count_inversions_recursive(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase , __lowerCamelCase : Dict = _count_cross_inversions(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : List[str] = inversion_p + inversions_q + cross_inversions return c, num_inversions def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Optional[int] = [] __lowerCamelCase : List[Any] = 0 while i < len(SCREAMING_SNAKE_CASE__ ) and j < len(SCREAMING_SNAKE_CASE__ ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(SCREAMING_SNAKE_CASE__ ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(SCREAMING_SNAKE_CASE__ ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def UpperCamelCase__ ( ): __lowerCamelCase : Optional[int] = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) __lowerCamelCase : Optional[Any] = count_inversions_bf(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase , __lowerCamelCase : Optional[Any] = count_inversions_recursive(SCREAMING_SNAKE_CASE__ ) assert num_inversions_bf == num_inversions_recursive == 8 print('number of inversions = ' , SCREAMING_SNAKE_CASE__ ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() __lowerCamelCase : Optional[Any] = count_inversions_bf(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase , __lowerCamelCase : int = count_inversions_recursive(SCREAMING_SNAKE_CASE__ ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , SCREAMING_SNAKE_CASE__ ) # an empty list should also have zero inversions __lowerCamelCase : List[str] = [] __lowerCamelCase : Dict = count_inversions_bf(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase , __lowerCamelCase : Dict = count_inversions_recursive(SCREAMING_SNAKE_CASE__ ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": main()
194
1
def _lowerCAmelCase ( __lowerCAmelCase ) -> tuple[int, int]: """simple docstring""" try: snake_case__ : int = float(__lowerCAmelCase ) except ValueError: raise ValueError('''Please enter a valid number''' ) snake_case__ : List[Any] = decimal - int(__lowerCAmelCase ) if fractional_part == 0: return int(__lowerCAmelCase ), 1 else: snake_case__ : List[Any] = len(str(__lowerCAmelCase ).split('''.''' )[1] ) snake_case__ : Dict = int(decimal * (10**number_of_frac_digits) ) snake_case__ : Tuple = 10**number_of_frac_digits snake_case__ , snake_case__ : Optional[int] = denominator, numerator while True: snake_case__ : List[Any] = dividend % divisor if remainder == 0: break snake_case__ , snake_case__ : List[str] = divisor, remainder snake_case__ , snake_case__ : Optional[int] = numerator / divisor, denominator / divisor return int(__lowerCAmelCase ), int(__lowerCAmelCase ) if __name__ == "__main__": print(f"""{decimal_to_fraction(2) = }""") print(f"""{decimal_to_fraction(89.0) = }""") print(f"""{decimal_to_fraction('67') = }""") print(f"""{decimal_to_fraction('45.0') = }""") print(f"""{decimal_to_fraction(1.5) = }""") print(f"""{decimal_to_fraction('6.25') = }""") print(f"""{decimal_to_fraction('78td') = }""")
230
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class a ( unittest.TestCase ): def __lowerCamelCase ( self :Union[str, Any] ): snake_case__ : Union[str, Any] = torch.tensor([-1_0_0, -1, -0.1, 0, 0.1, 1.0, 1_0_0] ) snake_case__ : int = get_activation('''gelu''' ) self.assertTrue(torch.allclose(gelu_python(__lowercase ) ,torch_builtin(__lowercase ) ) ) self.assertFalse(torch.allclose(gelu_python(__lowercase ) ,gelu_new(__lowercase ) ) ) def __lowerCamelCase ( self :Union[str, Any] ): snake_case__ : List[Any] = torch.tensor([-1_0_0, -1, -0.1, 0, 0.1, 1.0, 1_0_0] ) snake_case__ : Union[str, Any] = get_activation('''gelu''' ) snake_case__ : int = get_activation('''gelu_10''' ) snake_case__ : Optional[int] = torch_builtin(__lowercase ) snake_case__ : str = geluaa(__lowercase ) snake_case__ : Tuple = torch.where(y_gelu_aa < 10.0 ,1 ,0 ) self.assertTrue(torch.max(__lowercase ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask ,y_gelu_aa * clipped_mask ) ) def __lowerCamelCase ( self :Any ): get_activation('''gelu''' ) get_activation('''gelu_10''' ) get_activation('''gelu_fast''' ) get_activation('''gelu_new''' ) get_activation('''gelu_python''' ) get_activation('''gelu_pytorch_tanh''' ) get_activation('''linear''' ) get_activation('''mish''' ) get_activation('''quick_gelu''' ) get_activation('''relu''' ) get_activation('''sigmoid''' ) get_activation('''silu''' ) get_activation('''swish''' ) get_activation('''tanh''' ) with self.assertRaises(__lowercase ): get_activation('''bogus''' ) with self.assertRaises(__lowercase ): get_activation(__lowercase ) def __lowerCamelCase ( self :Optional[int] ): snake_case__ : str = get_activation('''gelu''' ) snake_case__ : List[Any] = 1 snake_case__ : Optional[Any] = get_activation('''gelu''' ) self.assertEqual(acta.a ,1 ) with self.assertRaises(__lowercase ): snake_case__ : str = acta.a
230
1
"""simple docstring""" import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig __snake_case : Union[str, Any] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE : def __init__( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = question_encoder lowerCAmelCase__ = generator lowerCAmelCase__ = self.question_encoder def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" if os.path.isfile(_snake_case ): raise ValueError(F"Provided path ({save_directory}) should be a directory, not a file" ) os.makedirs(_snake_case , exist_ok=_snake_case ) lowerCAmelCase__ = os.path.join(_snake_case , 'question_encoder_tokenizer' ) lowerCAmelCase__ = os.path.join(_snake_case , 'generator_tokenizer' ) self.question_encoder.save_pretrained(_snake_case ) self.generator.save_pretrained(_snake_case ) @classmethod def UpperCamelCase__ ( cls , _UpperCamelCase , **_UpperCamelCase ): """simple docstring""" from ..auto.tokenization_auto import AutoTokenizer lowerCAmelCase__ = kwargs.pop('config' , _snake_case ) if config is None: lowerCAmelCase__ = RagConfig.from_pretrained(_snake_case ) lowerCAmelCase__ = AutoTokenizer.from_pretrained( _snake_case , config=config.question_encoder , subfolder='question_encoder_tokenizer' ) lowerCAmelCase__ = AutoTokenizer.from_pretrained( _snake_case , config=config.generator , subfolder='generator_tokenizer' ) return cls(question_encoder=_snake_case , generator=_snake_case ) def __call__( self , *_UpperCamelCase , **_UpperCamelCase ): """simple docstring""" return self.current_tokenizer(*_snake_case , **_snake_case ) def UpperCamelCase__ ( self , *_UpperCamelCase , **_UpperCamelCase ): """simple docstring""" return self.generator.batch_decode(*_snake_case , **_snake_case ) def UpperCamelCase__ ( self , *_UpperCamelCase , **_UpperCamelCase ): """simple docstring""" return self.generator.decode(*_snake_case , **_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.question_encoder def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.generator def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = "longest" , _UpperCamelCase = None , _UpperCamelCase = True , **_UpperCamelCase , ): """simple docstring""" warnings.warn( '`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ' 'regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ' 'context manager to prepare your targets. See the documentation of your specific tokenizer for more ' 'details' , _snake_case , ) if max_length is None: lowerCAmelCase__ = self.current_tokenizer.model_max_length lowerCAmelCase__ = self( _snake_case , add_special_tokens=_snake_case , return_tensors=_snake_case , max_length=_snake_case , padding=_snake_case , truncation=_snake_case , **_snake_case , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: lowerCAmelCase__ = self.current_tokenizer.model_max_length lowerCAmelCase__ = self( text_target=_snake_case , add_special_tokens=_snake_case , return_tensors=_snake_case , padding=_snake_case , max_length=_snake_case , truncation=_snake_case , **_snake_case , ) lowerCAmelCase__ = labels['input_ids'] return model_inputs
357
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : Dict = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : str = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys __snake_case : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
122
0
def lowercase_ ( _A : int ): """simple docstring""" return str(_A ) == str(_A )[::-1] def lowercase_ ( _A : int ): """simple docstring""" return int(_A ) + int(str(_A )[::-1] ) def lowercase_ ( _A : int = 10000 ): """simple docstring""" lowerCamelCase__ : Optional[int] = [] for num in range(1 , _A ): lowerCamelCase__ : Union[str, Any] = 0 lowerCamelCase__ : Tuple = num while iterations < 50: lowerCamelCase__ : Optional[int] = sum_reverse(_A ) iterations += 1 if is_palindrome(_A ): break else: lychrel_nums.append(_A ) return len(_A ) if __name__ == "__main__": print(f'{solution() = }')
184
from collections import defaultdict def lowercase_ ( _A : int ): """simple docstring""" lowerCamelCase__ : Union[str, Any] = 1 lowerCamelCase__ : Dict = True for v in tree[start]: if v not in visited: ret += dfs(_A ) if ret % 2 == 0: cuts.append(_A ) return ret def lowercase_ ( ): """simple docstring""" dfs(1 ) if __name__ == "__main__": A, A : Tuple = 10, 9 A : int = defaultdict(list) A : dict[int, bool] = {} A : list[int] = [] A : List[str] = 0 A : Tuple = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
184
1
from __future__ import annotations import typing from collections import Counter def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(SCREAMING_SNAKE_CASE__ , max_perimeter + 1 ): __lowerCamelCase : int = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : int = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ = 1_000 ): __lowerCamelCase : int = pythagorean_triple(SCREAMING_SNAKE_CASE__ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(F"""Perimeter {solution()} has maximum solutions""")
194
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration lowercase_ = 5_0_0_0_0_0 lowercase_ ,lowercase_ = os.path.split(__file__) lowercase_ = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Optional[int] = dataset.map(**SCREAMING_SNAKE_CASE__ ) @get_duration def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Tuple = dataset.filter(**SCREAMING_SNAKE_CASE__ ) def UpperCamelCase__ ( ): __lowerCamelCase : str = {'num examples': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Any = datasets.Features({'text': datasets.Value('string' ), 'numbers': datasets.Value('float32' )} ) __lowerCamelCase : Any = generate_example_dataset( os.path.join(SCREAMING_SNAKE_CASE__ , 'dataset.arrow' ) , SCREAMING_SNAKE_CASE__ , num_examples=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : List[Any] = transformers.AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=SCREAMING_SNAKE_CASE__ ) def tokenize(SCREAMING_SNAKE_CASE__ ): return tokenizer(examples['text'] ) __lowerCamelCase : str = map(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : int = map(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[int] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='numpy' ): __lowerCamelCase : Union[str, Any] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='pandas' ): __lowerCamelCase : Any = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='torch' , columns='numbers' ): __lowerCamelCase : List[str] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type='tensorflow' , columns='numbers' ): __lowerCamelCase : List[Any] = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[Any] = map(SCREAMING_SNAKE_CASE__ , function=SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[int] = filter(SCREAMING_SNAKE_CASE__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(SCREAMING_SNAKE_CASE__ , 'wb' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ).encode('utf-8' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
194
1
"""simple docstring""" from collections import defaultdict def lowerCamelCase_ (UpperCamelCase__ : int ): _UpperCAmelCase : Union[str, Any] = 1 _UpperCAmelCase : Dict = True for v in tree[start]: if v not in visited: ret += dfs(UpperCamelCase_ ) if ret % 2 == 0: cuts.append(UpperCamelCase_ ) return ret def lowerCamelCase_ (): dfs(1 ) if __name__ == "__main__": _lowerCAmelCase,_lowerCAmelCase :str = 10, 9 _lowerCAmelCase :Any = defaultdict(list) _lowerCAmelCase :Optional[int] = {} _lowerCAmelCase :Union[str, Any] = [] _lowerCAmelCase :List[str] = 0 _lowerCAmelCase :List[Any] = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
263
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/config.json''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/config.json''', } class lowercase__ ( _UpperCAmelCase ): a_ ="""xlnet""" a_ =["""mems"""] a_ ={ """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __UpperCAmelCase=32000 , __UpperCAmelCase=1024 , __UpperCAmelCase=24 , __UpperCAmelCase=16 , __UpperCAmelCase=4096 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase="bi" , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=-1 , __UpperCAmelCase=False , __UpperCAmelCase="last" , __UpperCAmelCase=True , __UpperCAmelCase="tanh" , __UpperCAmelCase=0.1 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=1 , __UpperCAmelCase=2 , **__UpperCAmelCase , )-> int: '''simple docstring''' lowerCAmelCase__ = vocab_size lowerCAmelCase__ = d_model lowerCAmelCase__ = n_layer lowerCAmelCase__ = n_head if d_model % n_head != 0: raise ValueError(F"'d_model % n_head' ({d_model % n_head}) should be equal to 0" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F"`d_head` ({kwargs['d_head']}) should be equal to `d_model // n_head` ({d_model // n_head})" ) lowerCAmelCase__ = d_model // n_head lowerCAmelCase__ = ff_activation lowerCAmelCase__ = d_inner lowerCAmelCase__ = untie_r lowerCAmelCase__ = attn_type lowerCAmelCase__ = initializer_range lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = dropout lowerCAmelCase__ = mem_len lowerCAmelCase__ = reuse_len lowerCAmelCase__ = bi_data lowerCAmelCase__ = clamp_len lowerCAmelCase__ = same_length lowerCAmelCase__ = summary_type lowerCAmelCase__ = summary_use_proj lowerCAmelCase__ = summary_activation lowerCAmelCase__ = summary_last_dropout lowerCAmelCase__ = start_n_top lowerCAmelCase__ = end_n_top lowerCAmelCase__ = bos_token_id lowerCAmelCase__ = pad_token_id lowerCAmelCase__ = eos_token_id if "use_cache" in kwargs: warnings.warn( "The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`" " instead." , __UpperCAmelCase , ) lowerCAmelCase__ = kwargs["use_cache"] lowerCAmelCase__ = use_mems_eval lowerCAmelCase__ = use_mems_train super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) @property def UpperCAmelCase ( self )-> Dict: '''simple docstring''' logger.info(F"The model {self.model_type} is one of the few models that has no sequence length limit." ) return -1 @max_position_embeddings.setter def UpperCAmelCase ( self , __UpperCAmelCase )-> Union[str, Any]: '''simple docstring''' raise NotImplementedError( F"The model {self.model_type} is one of the few models that has no sequence length limit." )
340
0
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __lowerCamelCase : int = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = ["pixel_values"] def __init__( self : int , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = PILImageResampling.BILINEAR , _lowercase : bool = True , _lowercase : Union[int, float] = 1 / 2_55 , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : bool = True , **_lowercase : List[Any] , ): """simple docstring""" super().__init__(**_lowercase ) SCREAMING_SNAKE_CASE__ = size if size is not None else {"""shortest_edge""": 2_24} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else {"""height""": 2_56, """width""": 2_56} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size SCREAMING_SNAKE_CASE__ = resample SCREAMING_SNAKE_CASE__ = do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop SCREAMING_SNAKE_CASE__ = crop_size SCREAMING_SNAKE_CASE__ = do_flip_channel_order def __a ( self : List[Any] , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : PILImageResampling = PIL.Image.BILINEAR , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : str , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) SCREAMING_SNAKE_CASE__ = get_resize_output_image_size(_lowercase , size=size["""shortest_edge"""] , default_to_square=_lowercase ) return resize(_lowercase , size=_lowercase , resample=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : str , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowercase , size=(size["""height"""], size["""width"""]) , data_format=_lowercase , **_lowercase ) def __a ( self : Optional[Any] , _lowercase : np.ndarray , _lowercase : Union[int, float] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : Tuple , _lowercase : np.ndarray , _lowercase : Optional[Union[str, ChannelDimension]] = None ): """simple docstring""" return flip_channel_order(_lowercase , data_format=_lowercase ) def __a ( self : List[str] , _lowercase : ImageInput , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = None , _lowercase : bool = None , _lowercase : float = None , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : bool = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : ChannelDimension = ChannelDimension.FIRST , **_lowercase : int , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE__ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE__ = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) SCREAMING_SNAKE_CASE__ = size if size is not None else self.size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ = [to_numpy_array(_lowercase ) for image in images] if do_resize: SCREAMING_SNAKE_CASE__ = [self.resize(image=_lowercase , size=_lowercase , resample=_lowercase ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE__ = [self.center_crop(image=_lowercase , size=_lowercase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: SCREAMING_SNAKE_CASE__ = [self.flip_channel_order(image=_lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase ) def __a ( self : List[Any] , _lowercase : Dict , _lowercase : List[Tuple] = None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowercase ) != len(_lowercase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(_lowercase ): SCREAMING_SNAKE_CASE__ = target_sizes.numpy() SCREAMING_SNAKE_CASE__ = [] for idx in range(len(_lowercase ) ): SCREAMING_SNAKE_CASE__ = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=_lowercase ) SCREAMING_SNAKE_CASE__ = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowercase ) else: SCREAMING_SNAKE_CASE__ = logits.argmax(dim=1 ) SCREAMING_SNAKE_CASE__ = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
360
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 __lowerCamelCase : List[Any] = get_tests_dir('''fixtures''') __lowerCamelCase : Optional[int] = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') __lowerCamelCase : Any = get_tests_dir('''fixtures/dummy-config.json''') class __snake_case ( unittest.TestCase ): def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 0 def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""facebook/wav2vec2-base-960h""" ) self.assertIsInstance(_lowercase , _lowercase ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def __a ( self : Optional[int] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase ).to_dict() config_dict.pop("""feature_extractor_type""" ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor(**_lowercase ) # save in new folder model_config.save_pretrained(_lowercase ) config.save_pretrained(_lowercase ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase ) # make sure private variable is not incorrectly saved SCREAMING_SNAKE_CASE__ = json.loads(config.to_json_string() ) self.assertTrue("""_processor_class""" not in dict_as_saved ) self.assertIsInstance(_lowercase , _lowercase ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def __a ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( _lowercase , """bert-base is not a local folder and is not a valid model identifier""" ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""bert-base""" ) def __a ( self : Union[str, Any] ): """simple docstring""" with self.assertRaisesRegex( _lowercase , R"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase , revision="""aaaaaa""" ) def __a ( self : List[Any] ): """simple docstring""" with self.assertRaisesRegex( _lowercase , """hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.""" , ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""hf-internal-testing/config-no-model""" ) def __a ( self : str ): """simple docstring""" with self.assertRaises(_lowercase ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(_lowercase ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_lowercase ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase , trust_remote_code=_lowercase ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) def __a ( self : Union[str, Any] ): """simple docstring""" try: AutoConfig.register("""custom""" , _lowercase ) AutoFeatureExtractor.register(_lowercase , _lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_lowercase ): AutoFeatureExtractor.register(_lowercase , _lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API SCREAMING_SNAKE_CASE__ = CustomFeatureExtractor.from_pretrained(_lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_lowercase ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def __a ( self : Any ): """simple docstring""" class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = True try: AutoConfig.register("""custom""" , _lowercase ) AutoFeatureExtractor.register(_lowercase , _lowercase ) # If remote code is not set, the default is to use local SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(not hasattr(_lowercase , """is_local""" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
204
0
from ..utils import DummyObject, requires_backends class snake_case_ ( metaclass=__A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = ["note_seq"] def __init__( self : Optional[int] , *_UpperCamelCase : str , **_UpperCamelCase : Optional[int] ) ->Any: requires_backends(self , ['''note_seq'''] ) @classmethod def snake_case__( cls : int , *_UpperCamelCase : Any , **_UpperCamelCase : List[Any] ) ->int: requires_backends(cls , ['''note_seq'''] ) @classmethod def snake_case__( cls : Dict , *_UpperCamelCase : Optional[int] , **_UpperCamelCase : Union[str, Any] ) ->List[str]: requires_backends(cls , ['''note_seq'''] )
8
"""simple docstring""" import random class _UpperCAmelCase : @staticmethod def __snake_case ( _A ) -> tuple[list[int], list[int]]: '''simple docstring''' _UpperCAmelCase : List[Any] = [ord(_A ) for i in text] _UpperCAmelCase : str = [] _UpperCAmelCase : int = [] for i in plain: _UpperCAmelCase : List[str] = random.randint(1 , 3_00 ) _UpperCAmelCase : Any = (i + k) * k cipher.append(_A ) key.append(_A ) return cipher, key @staticmethod def __snake_case ( _A , _A ) -> str: '''simple docstring''' _UpperCAmelCase : Optional[int] = [] for i in range(len(_A ) ): _UpperCAmelCase : List[Any] = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(_A ) ) return "".join(_A ) if __name__ == "__main__": lowerCamelCase__ , lowerCamelCase__ : List[Any] = Onepad().encrypt('''Hello''') print(c, k) print(Onepad().decrypt(c, k))
246
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer UpperCamelCase__ = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast UpperCamelCase__ = TaTokenizerFast UpperCamelCase__ = {'configuration_mt5': ['MT5Config', 'MT5OnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'MT5EncoderModel', 'MT5ForConditionalGeneration', 'MT5ForQuestionAnswering', 'MT5Model', 'MT5PreTrainedModel', 'MT5Stack', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['TFMT5EncoderModel', 'TFMT5ForConditionalGeneration', 'TFMT5Model'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['FlaxMT5EncoderModel', 'FlaxMT5ForConditionalGeneration', 'FlaxMT5Model'] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys UpperCamelCase__ = _LazyModule( __name__, globals()['__file__'], _import_structure, extra_objects={'MT5Tokenizer': MTaTokenizer, 'MT5TokenizerFast': MTaTokenizerFast}, module_spec=__spec__, )
143
def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
143
1
"""simple docstring""" import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', } lowerCAmelCase_ = { 'vocab_file': {'ctrl': 'https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-vocab.json'}, 'merges_file': {'ctrl': 'https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-merges.txt'}, } lowerCAmelCase_ = { 'ctrl': 256, } lowerCAmelCase_ = { 'Pregnancy': 168_629, 'Christianity': 7_675, 'Explain': 106_423, 'Fitness': 63_440, 'Saving': 63_163, 'Ask': 27_171, 'Ass': 95_985, 'Joke': 163_509, 'Questions': 45_622, 'Thoughts': 49_605, 'Retail': 52_342, 'Feminism': 164_338, 'Writing': 11_992, 'Atheism': 192_263, 'Netflix': 48_616, 'Computing': 39_639, 'Opinion': 43_213, 'Alone': 44_967, 'Funny': 58_917, 'Gaming': 40_358, 'Human': 4_088, 'India': 1_331, 'Joker': 77_138, 'Diet': 36_206, 'Legal': 11_859, 'Norman': 4_939, 'Tip': 72_689, 'Weight': 52_343, 'Movies': 46_273, 'Running': 23_425, 'Science': 2_090, 'Horror': 37_793, 'Confession': 60_572, 'Finance': 12_250, 'Politics': 16_360, 'Scary': 191_985, 'Support': 12_654, 'Technologies': 32_516, 'Teenage': 66_160, 'Event': 32_769, 'Learned': 67_460, 'Notion': 182_770, 'Wikipedia': 37_583, 'Books': 6_665, 'Extract': 76_050, 'Confessions': 102_701, 'Conspiracy': 75_932, 'Links': 63_674, 'Narcissus': 150_425, 'Relationship': 54_766, 'Relationships': 134_796, 'Reviews': 41_671, 'News': 4_256, 'Translation': 26_820, 'multilingual': 128_406, } def __UpperCAmelCase ( __lowerCamelCase ) -> Union[str, Any]: lowercase__ : Dict = set() lowercase__ : Optional[int] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Dict = char lowercase__ : Tuple = set(__lowerCamelCase ) return pairs class __A ( A_ ): '''simple docstring''' lowerCAmelCase : List[str] = VOCAB_FILES_NAMES lowerCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase : Optional[Any] = CONTROL_CODES def __init__( self : int ,_snake_case : str ,_snake_case : Tuple ,_snake_case : List[str]="<unk>" ,**_snake_case : List[str] ) -> Tuple: """simple docstring""" super().__init__(unk_token=_snake_case ,**_snake_case ) with open(_snake_case ,encoding='''utf-8''' ) as vocab_handle: lowercase__ : Dict = json.load(_snake_case ) lowercase__ : str = {v: k for k, v in self.encoder.items()} with open(_snake_case ,encoding='''utf-8''' ) as merges_handle: lowercase__ : Union[str, Any] = merges_handle.read().split('''\n''' )[1:-1] lowercase__ : List[Any] = [tuple(merge.split() ) for merge in merges] lowercase__ : List[str] = dict(zip(_snake_case ,range(len(_snake_case ) ) ) ) lowercase__ : int = {} @property def UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" return len(self.encoder ) def UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase ( self : Any ,_snake_case : Union[str, Any] ) -> Tuple: """simple docstring""" if token in self.cache: return self.cache[token] lowercase__ : str = tuple(_snake_case ) lowercase__ : Dict = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowercase__ : Any = get_pairs(_snake_case ) if not pairs: return token while True: lowercase__ : Dict = min(_snake_case ,key=lambda _snake_case : self.bpe_ranks.get(_snake_case ,float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : Any = bigram lowercase__ : Tuple = [] lowercase__ : Any = 0 while i < len(_snake_case ): try: lowercase__ : Optional[Any] = word.index(_snake_case ,_snake_case ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : int = j if word[i] == first and i < len(_snake_case ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : Tuple = tuple(_snake_case ) lowercase__ : int = new_word if len(_snake_case ) == 1: break else: lowercase__ : Optional[Any] = get_pairs(_snake_case ) lowercase__ : Union[str, Any] = '''@@ '''.join(_snake_case ) lowercase__ : List[Any] = word[:-4] lowercase__ : int = word return word def UpperCAmelCase ( self : Optional[Any] ,_snake_case : Optional[int] ) -> Optional[int]: """simple docstring""" lowercase__ : List[str] = [] lowercase__ : int = re.findall(r'''\S+\n?''' ,_snake_case ) for token in words: split_tokens.extend(list(self.bpe(_snake_case ).split(''' ''' ) ) ) return split_tokens def UpperCAmelCase ( self : List[Any] ,_snake_case : List[Any] ) -> Tuple: """simple docstring""" return self.encoder.get(_snake_case ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase ( self : Optional[int] ,_snake_case : str ) -> Optional[Any]: """simple docstring""" return self.decoder.get(_snake_case ,self.unk_token ) def UpperCAmelCase ( self : List[str] ,_snake_case : str ) -> Any: """simple docstring""" lowercase__ : List[str] = ''' '''.join(_snake_case ).replace('''@@ ''' ,'''''' ).strip() return out_string def UpperCAmelCase ( self : Dict ,_snake_case : str ,_snake_case : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_snake_case ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : int = os.path.join( _snake_case ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase__ : Dict = os.path.join( _snake_case ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(_snake_case ,'''w''' ,encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=_snake_case ,ensure_ascii=_snake_case ) + '''\n''' ) lowercase__ : Optional[int] = 0 with open(_snake_case ,'''w''' ,encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda _snake_case : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) lowercase__ : List[Any] = token_index writer.write(''' '''.join(_snake_case ) + '''\n''' ) index += 1 return vocab_file, merge_file # def decode(self, token_ids, skip_special_tokens=False, clean_up_tokenization_spaces=True): # filtered_tokens = ' '.join(self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)) # tokens_generated_so_far = re.sub('(@@ )', '', string=filtered_tokens) # tokens_generated_so_far = re.sub('(@@ ?$)', '', string=tokens_generated_so_far) # return ''.join(tokens_generated_so_far)
16
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : str = "" ): __UpperCamelCase =url or 'https://www.imdb.com/chart/top/?ref_=nv_mv_250' __UpperCamelCase =BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE__ ).text , 'html.parser' ) __UpperCamelCase =soup.find_all('td' , attrs='titleColumn' ) __UpperCamelCase =soup.find_all('td' , class_='ratingColumn imdbRating' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) } def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : str = "IMDb_Top_250_Movies.csv" ): __UpperCamelCase =get_imdb_top_aaa_movies() with open(SCREAMING_SNAKE_CASE__ , 'w' , newline='' ) as out_file: __UpperCamelCase =csv.writer(SCREAMING_SNAKE_CASE__ ) writer.writerow(['Movie title', 'IMDb rating'] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
62
0
import numpy as np import qiskit def lowerCamelCase ( UpperCAmelCase__ : int = 8 , UpperCAmelCase__ : int | None = None ) -> str: lowercase_ : Tuple = np.random.default_rng(seed=UpperCAmelCase__ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. lowercase_ : Dict = 6 * key_len # Measurement basis for Alice's qubits. lowercase_ : int = rng.integers(2 , size=UpperCAmelCase__ ) # The set of states Alice will prepare. lowercase_ : Dict = rng.integers(2 , size=UpperCAmelCase__ ) # Measurement basis for Bob's qubits. lowercase_ : Tuple = rng.integers(2 , size=UpperCAmelCase__ ) # Quantum Circuit to simulate BB84 lowercase_ : int = qiskit.QuantumCircuit(UpperCAmelCase__ , name="""BB84""" ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(UpperCAmelCase__ ): if alice_state[index] == 1: bbaa_circ.x(UpperCAmelCase__ ) if alice_basis[index] == 1: bbaa_circ.h(UpperCAmelCase__ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(UpperCAmelCase__ ): if bob_basis[index] == 1: bbaa_circ.h(UpperCAmelCase__ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. lowercase_ : int = qiskit.Aer.get_backend("""aer_simulator""" ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. lowercase_ : List[str] = qiskit.execute(UpperCAmelCase__ , UpperCAmelCase__ , shots=1 , seed_simulator=UpperCAmelCase__ ) # Returns the result of measurement. lowercase_ : Any = job.result().get_counts(UpperCAmelCase__ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. lowercase_ : Union[str, Any] = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. lowercase_ : Union[str, Any] = gen_key[:key_len] if len(UpperCAmelCase__ ) >= key_len else gen_key.ljust(UpperCAmelCase__ , """0""" ) return key if __name__ == "__main__": print(f"""The generated key is : {bbaa(8, seed=0)}""") from doctest import testmod testmod()
371
'''simple docstring''' import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class __magic_name__ ( unittest.TestCase): @parameterized.expand([(None,), ("""foo.json""",)] ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] , lowercase_ : str ): lowercase_ : Union[str, Any] = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowercase_ , config_name=lowercase_ ) lowercase_ : Optional[int] = GenerationConfig.from_pretrained(lowercase_ , config_name=lowercase_ ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , lowercase_ ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 50 ) self.assertEqual(loaded_config.max_length , 20 ) self.assertEqual(loaded_config.max_time , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : int = AutoConfig.from_pretrained("""gpt2""" ) lowercase_ : List[Any] = GenerationConfig.from_model_config(lowercase_ ) lowercase_ : Optional[int] = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(lowercase_ , lowercase_ ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def SCREAMING_SNAKE_CASE_ ( self : List[str] ): lowercase_ : Optional[int] = GenerationConfig() lowercase_ : int = { """max_new_tokens""": 1024, """foo""": """bar""", } lowercase_ : List[str] = copy.deepcopy(lowercase_ ) lowercase_ : Tuple = generation_config.update(**lowercase_ ) # update_kwargs was not modified (no side effects) self.assertEqual(lowercase_ , lowercase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 1024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(lowercase_ , {"""foo""": """bar"""} ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Dict = GenerationConfig() lowercase_ : int = """bar""" with tempfile.TemporaryDirectory("""test-generation-config""" ) as tmp_dir: generation_config.save_pretrained(lowercase_ ) lowercase_ : Optional[int] = GenerationConfig.from_pretrained(lowercase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , """bar""" ) lowercase_ : List[str] = GenerationConfig.from_model_config(lowercase_ ) assert not hasattr(lowercase_ , """foo""" ) # no new kwargs should be initialized if from config def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Optional[int] = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , lowercase_ ) self.assertEqual(default_config.num_beams , 1 ) lowercase_ : Dict = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , lowercase_ ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowercase_ ) lowercase_ : Tuple = GenerationConfig.from_pretrained(lowercase_ , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , lowercase_ ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class __magic_name__ ( unittest.TestCase): @classmethod def SCREAMING_SNAKE_CASE_ ( cls : Any ): lowercase_ : int = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def SCREAMING_SNAKE_CASE_ ( cls : List[Any] ): try: delete_repo(token=cls._token , repo_id="""test-generation-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-generation-config-org""" ) except HTTPError: pass def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : Tuple = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""test-generation-config""" , use_auth_token=self._token ) lowercase_ : List[Any] = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-generation-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowercase_ , repo_id="""test-generation-config""" , push_to_hub=lowercase_ , use_auth_token=self._token ) lowercase_ : int = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : List[Any] = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""valid_org/test-generation-config-org""" , use_auth_token=self._token ) lowercase_ : Optional[Any] = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-generation-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowercase_ , repo_id="""valid_org/test-generation-config-org""" , push_to_hub=lowercase_ , use_auth_token=self._token ) lowercase_ : int = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) )
21
0
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) __lowerCAmelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name __lowerCAmelCase : Any = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline\n >>> from diffusers.utils import load_image\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior.to("cuda")\n\n >>> prompt = "A red cartoon frog, 4k"\n >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False)\n\n >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16\n ... )\n >>> pipe.to("cuda")\n\n >>> init_image = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/frog.png"\n ... )\n\n >>> image = pipe(\n ... image=init_image,\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... strength=0.2,\n ... ).images\n\n >>> image[0].save("red_frog.png")\n ```\n' def a__ ( A_, A_, A_=8 ): '''simple docstring''' __magic_name__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __magic_name__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def a__ ( A_, A_=512, A_=512 ): '''simple docstring''' __magic_name__ = pil_image.resize((w, h), resample=Image.BICUBIC, reducing_gap=1 ) __magic_name__ = np.array(pil_image.convert("""RGB""" ) ) __magic_name__ = arr.astype(np.floataa ) / 127.5 - 1 __magic_name__ = np.transpose(A_, [2, 0, 1] ) __magic_name__ = torch.from_numpy(A_ ).unsqueeze(0 ) return image class UpperCAmelCase_ ( _A ): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : UNetaDConditionModel , UpperCamelCase__ : DDPMScheduler , UpperCamelCase__ : VQModel , ) -> int: """simple docstring""" super().__init__() self.register_modules( unet=UpperCamelCase__ , scheduler=UpperCamelCase__ , movq=UpperCamelCase__ , ) __magic_name__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] ) -> str: """simple docstring""" __magic_name__ = min(int(num_inference_steps * strength ) , UpperCamelCase__ ) __magic_name__ = max(num_inference_steps - init_timestep , 0 ) __magic_name__ = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def _lowercase ( self : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple=None ) -> Union[str, Any]: """simple docstring""" if not isinstance(UpperCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(UpperCamelCase__ )}''' ) __magic_name__ = image.to(device=UpperCamelCase__ , dtype=UpperCamelCase__ ) __magic_name__ = batch_size * num_images_per_prompt if image.shape[1] == 4: __magic_name__ = image else: if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(UpperCamelCase__ )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): __magic_name__ = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(UpperCamelCase__ ) ] __magic_name__ = torch.cat(UpperCamelCase__ , dim=0 ) else: __magic_name__ = self.movq.encode(UpperCamelCase__ ).latent_dist.sample(UpperCamelCase__ ) __magic_name__ = self.movq.config.scaling_factor * init_latents __magic_name__ = torch.cat([init_latents] , dim=0 ) __magic_name__ = init_latents.shape __magic_name__ = randn_tensor(UpperCamelCase__ , generator=UpperCamelCase__ , device=UpperCamelCase__ , dtype=UpperCamelCase__ ) # get latents __magic_name__ = self.scheduler.add_noise(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = init_latents return latents def _lowercase ( self : Tuple , UpperCamelCase__ : Tuple=0 ) -> List[Any]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) __magic_name__ = torch.device(F'''cuda:{gpu_id}''' ) __magic_name__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self : Tuple , UpperCamelCase__ : Any=0 ) -> str: """simple docstring""" if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) __magic_name__ = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=UpperCamelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __magic_name__ = None for cpu_offloaded_model in [self.unet, self.movq]: __magic_name__ , __magic_name__ = cpu_offload_with_hook(UpperCamelCase__ , UpperCamelCase__ , prev_module_hook=UpperCamelCase__ ) # We'll offload the last model manually. __magic_name__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _lowercase ( self : List[str] ) -> List[str]: """simple docstring""" if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCamelCase__ , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCamelCase__ ) def __call__( self : Optional[int] , UpperCamelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image, List[torch.FloatTensor], List[PIL.Image.Image]] , UpperCamelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCamelCase__ : int = 512 , UpperCamelCase__ : int = 512 , UpperCamelCase__ : int = 100 , UpperCamelCase__ : float = 4.0 , UpperCamelCase__ : float = 0.3 , UpperCamelCase__ : int = 1 , UpperCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCamelCase__ : Optional[str] = "pil" , UpperCamelCase__ : bool = True , ) -> int: """simple docstring""" __magic_name__ = self._execution_device __magic_name__ = guidance_scale > 1.0 if isinstance(UpperCamelCase__ , UpperCamelCase__ ): __magic_name__ = torch.cat(UpperCamelCase__ , dim=0 ) __magic_name__ = image_embeds.shape[0] if isinstance(UpperCamelCase__ , UpperCamelCase__ ): __magic_name__ = torch.cat(UpperCamelCase__ , dim=0 ) if do_classifier_free_guidance: __magic_name__ = image_embeds.repeat_interleave(UpperCamelCase__ , dim=0 ) __magic_name__ = negative_image_embeds.repeat_interleave(UpperCamelCase__ , dim=0 ) __magic_name__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCamelCase__ ) if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): __magic_name__ = [image] if not all(isinstance(UpperCamelCase__ , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F'''Input is in incorrect format: {[type(UpperCamelCase__ ) for i in image]}. Currently, we only support PIL image and pytorch tensor''' ) __magic_name__ = torch.cat([prepare_image(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) for i in image] , dim=0 ) __magic_name__ = image.to(dtype=image_embeds.dtype , device=UpperCamelCase__ ) __magic_name__ = self.movq.encode(UpperCamelCase__ )["""latents"""] __magic_name__ = latents.repeat_interleave(UpperCamelCase__ , dim=0 ) self.scheduler.set_timesteps(UpperCamelCase__ , device=UpperCamelCase__ ) __magic_name__ , __magic_name__ = self.get_timesteps(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __magic_name__ , __magic_name__ = downscale_height_and_width(UpperCamelCase__ , UpperCamelCase__ , self.movq_scale_factor ) __magic_name__ = self.prepare_latents( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , image_embeds.dtype , UpperCamelCase__ , UpperCamelCase__ ) for i, t in enumerate(self.progress_bar(UpperCamelCase__ ) ): # expand the latents if we are doing classifier free guidance __magic_name__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __magic_name__ = {"""image_embeds""": image_embeds} __magic_name__ = self.unet( sample=UpperCamelCase__ , timestep=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , added_cond_kwargs=UpperCamelCase__ , return_dict=UpperCamelCase__ , )[0] if do_classifier_free_guidance: __magic_name__ , __magic_name__ = noise_pred.split(latents.shape[1] , dim=1 ) __magic_name__ , __magic_name__ = noise_pred.chunk(2 ) __magic_name__ , __magic_name__ = variance_pred.chunk(2 ) __magic_name__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __magic_name__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __magic_name__ , __magic_name__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __magic_name__ = self.scheduler.step( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ , )[0] # post-processing __magic_name__ = self.movq.decode(UpperCamelCase__ , force_not_quantize=UpperCamelCase__ )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: __magic_name__ = image * 0.5 + 0.5 __magic_name__ = image.clamp(0 , 1 ) __magic_name__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __magic_name__ = self.numpy_to_pil(UpperCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCamelCase__ )
88
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys __lowerCAmelCase : int = subprocess.check_output('git merge-base main HEAD'.split()).decode('utf-8') __lowerCAmelCase : Any = ( subprocess.check_output(F'''git diff --diff-filter=d --name-only {fork_point_sha}'''.split()).decode('utf-8').split() ) __lowerCAmelCase : str = '|'.join(sys.argv[1:]) __lowerCAmelCase : Tuple = re.compile(RF'''^({joined_dirs}).*?\.py$''') __lowerCAmelCase : Union[str, Any] = [x for x in modified_files if regex.match(x)] print(' '.join(relevant_modified_files), end='')
88
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _snake_case : def __init__( self , _a , _a=13 , _a=2 , _a=24 , _a=16 , _a=True , _a=True , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=10 , _a=0.02 , _a=None , _a=2 , _a=2 , ): __magic_name__ : Union[str, Any] = parent __magic_name__ : str = batch_size __magic_name__ : List[str] = patch_size __magic_name__ : Tuple = max_length __magic_name__ : List[str] = num_mel_bins __magic_name__ : Union[str, Any] = is_training __magic_name__ : Union[str, Any] = use_labels __magic_name__ : str = hidden_size __magic_name__ : Tuple = num_hidden_layers __magic_name__ : Union[str, Any] = num_attention_heads __magic_name__ : int = intermediate_size __magic_name__ : Optional[int] = hidden_act __magic_name__ : Union[str, Any] = hidden_dropout_prob __magic_name__ : Dict = attention_probs_dropout_prob __magic_name__ : Optional[Any] = type_sequence_label_size __magic_name__ : Union[str, Any] = initializer_range __magic_name__ : List[str] = scope __magic_name__ : int = frequency_stride __magic_name__ : Any = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) __magic_name__ : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 __magic_name__ : str = (self.max_length - self.patch_size) // self.time_stride + 1 __magic_name__ : Optional[Any] = frequency_out_dimension * time_out_dimension __magic_name__ : Optional[int] = num_patches + 2 def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) __magic_name__ : List[Any] = None if self.use_labels: __magic_name__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ : Union[str, Any] = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a ): __magic_name__ : Optional[int] = ASTModel(config=_a ) model.to(_a ) model.eval() __magic_name__ : Tuple = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = self.prepare_config_and_inputs() ( ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ) : Optional[int] = config_and_inputs __magic_name__ : Optional[int] = {"input_values": input_values} return config, inputs_dict @require_torch class _snake_case ( snake_case , snake_case , unittest.TestCase ): UpperCamelCase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) UpperCamelCase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def SCREAMING_SNAKE_CASE ( self , _a , _a , _a , _a , _a ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[Any] = ASTModelTester(self ) __magic_name__ : int = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def SCREAMING_SNAKE_CASE ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="AST does not use inputs_embeds" ) def SCREAMING_SNAKE_CASE ( self ): pass def SCREAMING_SNAKE_CASE ( self ): __magic_name__ , __magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ : List[str] = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __magic_name__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ , __magic_name__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ : Optional[int] = model_class(_a ) __magic_name__ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __magic_name__ : str = [*signature.parameters.keys()] __magic_name__ : Dict = ["input_values"] self.assertListEqual(arg_names[:1] , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) @slow def SCREAMING_SNAKE_CASE ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __magic_name__ : Optional[int] = ASTModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' __magic_name__ : Dict = hf_hub_download( repo_id="nielsr/audio-spectogram-transformer-checkpoint" , filename="sample_audio.flac" , repo_type="dataset" ) __magic_name__ , __magic_name__ : Tuple = torchaudio.load(_snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _snake_case ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE ( self ): return ( ASTFeatureExtractor.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.default_feature_extractor __magic_name__ : Tuple = ASTForAudioClassification.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ).to(_a ) __magic_name__ : Any = self.default_feature_extractor __magic_name__ , __magic_name__ : int = prepare_audio() __magic_name__ : Dict = audio.squeeze().numpy() __magic_name__ : Dict = feature_extractor(_a , sampling_rate=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): __magic_name__ : Tuple = model(**_a ) # verify the logits __magic_name__ : Optional[Any] = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , _a ) __magic_name__ : Union[str, Any] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1e-4 ) )
41
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Union[str, Any] = logging.get_logger(__name__) snake_case : List[Any] = { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _snake_case ( snake_case ): UpperCamelCase__ = 'roformer' def __init__( self , _a=50_000 , _a=None , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=1_536 , _a=2 , _a=0.02 , _a=1e-12 , _a=0 , _a=False , _a=True , **_a , ): super().__init__(pad_token_id=_a , **_a ) __magic_name__ : Tuple = vocab_size __magic_name__ : Dict = hidden_size if embedding_size is None else embedding_size __magic_name__ : int = hidden_size __magic_name__ : int = num_hidden_layers __magic_name__ : Union[str, Any] = num_attention_heads __magic_name__ : Union[str, Any] = hidden_act __magic_name__ : Optional[int] = intermediate_size __magic_name__ : Union[str, Any] = hidden_dropout_prob __magic_name__ : Union[str, Any] = attention_probs_dropout_prob __magic_name__ : Tuple = max_position_embeddings __magic_name__ : str = type_vocab_size __magic_name__ : Dict = initializer_range __magic_name__ : Tuple = layer_norm_eps __magic_name__ : Optional[int] = rotary_value __magic_name__ : List[Any] = use_cache class _snake_case ( snake_case ): @property def SCREAMING_SNAKE_CASE ( self ): if self.task == "multiple-choice": __magic_name__ : str = {0: "batch", 1: "choice", 2: "sequence"} else: __magic_name__ : str = {0: "batch", 1: "sequence"} __magic_name__ : Tuple = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
41
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """facebook/timesformer""": """https://huggingface.co/facebook/timesformer/resolve/main/config.json""", } class UpperCAmelCase_ ( lowercase ): """simple docstring""" UpperCamelCase_ : List[Any] ='timesformer' def __init__( self , SCREAMING_SNAKE_CASE_=224 , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=8 , SCREAMING_SNAKE_CASE_=768 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=3072 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.0 , SCREAMING_SNAKE_CASE_=0.0 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=1e-6 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_="divided_space_time" , SCREAMING_SNAKE_CASE_=0 , **SCREAMING_SNAKE_CASE_ , ) -> Tuple: super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :str = image_size UpperCamelCase :Optional[int] = patch_size UpperCamelCase :Tuple = num_channels UpperCamelCase :Union[str, Any] = num_frames UpperCamelCase :Tuple = hidden_size UpperCamelCase :Optional[Any] = num_hidden_layers UpperCamelCase :Optional[int] = num_attention_heads UpperCamelCase :int = intermediate_size UpperCamelCase :Dict = hidden_act UpperCamelCase :int = hidden_dropout_prob UpperCamelCase :Any = attention_probs_dropout_prob UpperCamelCase :List[Any] = initializer_range UpperCamelCase :Tuple = layer_norm_eps UpperCamelCase :List[str] = qkv_bias UpperCamelCase :str = attention_type UpperCamelCase :Any = drop_path_rate
259
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, is_batched, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( lowercase ): """simple docstring""" UpperCamelCase_ : List[Any] =['pixel_values'] def __init__( self , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = PILImageResampling.BICUBIC , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = 1 / 255 , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> None: super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Dict = size if size is not None else {'''height''': 224, '''width''': 224} UpperCamelCase :Optional[Any] = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :List[Any] = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} UpperCamelCase :Dict = get_size_dict(SCREAMING_SNAKE_CASE_ , default_to_square=SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' ) UpperCamelCase :Optional[int] = do_resize UpperCamelCase :int = do_rescale UpperCamelCase :Tuple = do_normalize UpperCamelCase :str = do_center_crop UpperCamelCase :int = crop_size UpperCamelCase :Tuple = size UpperCamelCase :List[str] = resample UpperCamelCase :Tuple = rescale_factor UpperCamelCase :Optional[Any] = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN UpperCamelCase :Optional[int] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = PILImageResampling.BILINEAR , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: UpperCamelCase :Dict = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "shortest_edge" in size: UpperCamelCase :str = get_resize_output_image_size(SCREAMING_SNAKE_CASE_ , size=size['''shortest_edge'''] , default_to_square=SCREAMING_SNAKE_CASE_ ) # size = get_resize_output_image_size(image, size["shortest_edge"], size["longest_edge"]) elif "height" in size and "width" in size: UpperCamelCase :Optional[int] = (size['''height'''], size['''width''']) else: raise ValueError(F'''Size must contain \'height\' and \'width\' keys or \'shortest_edge\' key. Got {size.keys()}''' ) return resize(SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: UpperCamelCase :Union[str, Any] = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(SCREAMING_SNAKE_CASE_ , size=(size['''height'''], size['''width''']) , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ ) -> np.ndarray: return rescale(SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: return normalize(SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE_ , ) -> BatchFeature: UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize UpperCamelCase :Optional[int] = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase :Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase :Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase :Optional[int] = crop_size if crop_size is not None else self.crop_size UpperCamelCase :Dict = get_size_dict(SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' , default_to_square=SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Dict = resample if resample is not None else self.resample UpperCamelCase :List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase :Optional[Any] = image_mean if image_mean is not None else self.image_mean UpperCamelCase :Dict = image_std if image_std is not None else self.image_std UpperCamelCase :Dict = size if size is not None else self.size UpperCamelCase :Optional[int] = get_size_dict(SCREAMING_SNAKE_CASE_ ) if not is_batched(SCREAMING_SNAKE_CASE_ ): UpperCamelCase :str = [images] if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. UpperCamelCase :Tuple = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] if do_resize: UpperCamelCase :List[Any] = [self.resize(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_center_crop: UpperCamelCase :Tuple = [self.center_crop(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: UpperCamelCase :Union[str, Any] = [self.rescale(image=SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: UpperCamelCase :Union[str, Any] = [self.normalize(image=SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase :List[str] = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=SCREAMING_SNAKE_CASE_ , tensor_type=SCREAMING_SNAKE_CASE_ )
259
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A = logging.get_logger(__name__) __A = {'vocab_file': 'spiece.model'} __A = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="<s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="<sep>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="<cls>" , __UpperCAmelCase="<mask>" , __UpperCAmelCase=["<eop>", "<eod>"] , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> None: _lowerCAmelCase =AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token _lowerCAmelCase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) _lowerCAmelCase =3 _lowerCAmelCase =do_lower_case _lowerCAmelCase =remove_space _lowerCAmelCase =keep_accents _lowerCAmelCase =vocab_file _lowerCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) _lowerCAmelCase =jieba _lowerCAmelCase =str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _lowerCAmelCase ( self ) -> Dict: return len(self.sp_model ) def _lowerCAmelCase ( self ) -> Any: _lowerCAmelCase ={self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Any: _lowerCAmelCase =self.__dict__.copy() _lowerCAmelCase =None return state def __setstate__( self , __UpperCAmelCase ) -> List[Any]: _lowerCAmelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _lowerCAmelCase ={} _lowerCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowerCAmelCase ( self , __UpperCAmelCase ) -> str: if self.remove_space: _lowerCAmelCase =""" """.join(inputs.strip().split() ) else: _lowerCAmelCase =inputs _lowerCAmelCase =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: _lowerCAmelCase =unicodedata.normalize("""NFKD""" , __UpperCAmelCase ) _lowerCAmelCase ="""""".join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: _lowerCAmelCase =outputs.lower() return outputs def _lowerCAmelCase ( self , __UpperCAmelCase ) -> List[str]: _lowerCAmelCase =self.preprocess_text(__UpperCAmelCase ) _lowerCAmelCase =self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) _lowerCAmelCase =[] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): _lowerCAmelCase =self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _lowerCAmelCase =cur_pieces[1:] else: _lowerCAmelCase =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def _lowerCAmelCase ( self , __UpperCAmelCase ) -> str: return self.sp_model.PieceToId(__UpperCAmelCase ) def _lowerCAmelCase ( self , __UpperCAmelCase ) -> List[str]: return self.sp_model.IdToPiece(__UpperCAmelCase ) def _lowerCAmelCase ( self , __UpperCAmelCase ) -> int: _lowerCAmelCase ="""""".join(__UpperCAmelCase ).replace(__UpperCAmelCase , """ """ ).strip() return out_string def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> List[int]: _lowerCAmelCase =[self.sep_token_id] _lowerCAmelCase =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] return ([0] * len(__UpperCAmelCase )) + [1, 1] def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> List[int]: _lowerCAmelCase =[self.sep_token_id] _lowerCAmelCase =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> Tuple[str]: if not os.path.isdir(__UpperCAmelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return _lowerCAmelCase =os.path.join( __UpperCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , """wb""" ) as fi: _lowerCAmelCase =self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,) def _lowerCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: _lowerCAmelCase =super()._decode(*__UpperCAmelCase , **__UpperCAmelCase ) _lowerCAmelCase =text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
353
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ['PerceiverFeatureExtractor'] __A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = StableUnCLIPPipeline lowercase__ = TEXT_TO_IMAGE_PARAMS lowercase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowercase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowercase__ = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false lowercase__ = False def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = 32 _UpperCamelCase = embedder_hidden_size # prior components torch.manual_seed(0) _UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) _UpperCamelCase = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__a , projection_dim=__a , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , )) torch.manual_seed(0) _UpperCamelCase = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=__a , num_layers=1 , ) torch.manual_seed(0) _UpperCamelCase = DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=10_00 , clip_sample=__a , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0) _UpperCamelCase = StableUnCLIPImageNormalizer(embedding_dim=__a) _UpperCamelCase = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) _UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) _UpperCamelCase = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__a , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , )) torch.manual_seed(0) _UpperCamelCase = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__a , layers_per_block=1 , upcast_attention=__a , use_linear_projection=__a , ) torch.manual_seed(0) _UpperCamelCase = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_0085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=__a , steps_offset=1 , ) torch.manual_seed(0) _UpperCamelCase = AutoencoderKL() _UpperCamelCase = { # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def UpperCAmelCase ( self , __a , __a=0) -> Tuple: '''simple docstring''' if str(__a).startswith('''mps'''): _UpperCamelCase = torch.manual_seed(__a) else: _UpperCamelCase = torch.Generator(device=__a).manual_seed(__a) _UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__a) @slow @require_torch_gpu class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''') _UpperCamelCase = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa) pipe.to(__a) pipe.set_progress_bar_config(disable=__a) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() _UpperCamelCase = torch.Generator(device='''cpu''').manual_seed(0) _UpperCamelCase = pipe('''anime turle''' , generator=__a , output_type='''np''') _UpperCamelCase = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(__a , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _UpperCamelCase = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa) _UpperCamelCase = pipe.to(__a) pipe.set_progress_bar_config(disable=__a) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() _UpperCamelCase = pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) _UpperCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
194
"""simple docstring""" from sklearn.metrics import matthews_corrcoef import datasets _a = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ _a = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ _a = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase( datasets.Metric ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int32'''), '''references''': datasets.Value('''int32'''), }) , reference_urls=[ '''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html''' ] , ) def UpperCAmelCase ( self , __a , __a , __a=None) -> Dict: '''simple docstring''' return { "matthews_correlation": float(matthews_corrcoef(__a , __a , sample_weight=__a)), }
194
1
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser _snake_case : Optional[int] = logging.getLogger(__name__) torch.set_grad_enabled(False) _snake_case : Union[str, Any] = 'cuda' if torch.cuda.is_available() else 'cpu' def snake_case_ (UpperCamelCase : str , UpperCamelCase : Any=100 , UpperCamelCase : Optional[int]=" " ): '''simple docstring''' _a = text.split(UpperCamelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(UpperCamelCase ) , UpperCamelCase )] def snake_case_ (UpperCamelCase : dict ): '''simple docstring''' _a , _a = [], [] for title, text in zip(documents['''title'''] , documents['''text'''] ): if text is not None: for passage in split_text(UpperCamelCase ): titles.append(title if title is not None else '''''' ) texts.append(UpperCamelCase ) return {"title": titles, "text": texts} def snake_case_ (UpperCamelCase : dict , UpperCamelCase : DPRContextEncoder , UpperCamelCase : DPRContextEncoderTokenizerFast ): '''simple docstring''' _a = ctx_tokenizer( documents['''title'''] , documents['''text'''] , truncation=UpperCamelCase , padding='''longest''' , return_tensors='''pt''' )['''input_ids'''] _a = ctx_encoder(input_ids.to(device=UpperCamelCase ) , return_dict=UpperCamelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def snake_case_ (UpperCamelCase : "RagExampleArguments" , UpperCamelCase : "ProcessingArguments" , UpperCamelCase : "IndexHnswArguments" , ): '''simple docstring''' logger.info('''Step 1 - Create the dataset''' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _a = load_dataset( '''csv''' , data_files=[rag_example_args.csv_path] , split='''train''' , delimiter='''\t''' , column_names=['''title''', '''text'''] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _a = dataset.map(UpperCamelCase , batched=UpperCamelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _a = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=UpperCamelCase ) _a = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _a = Features( {'''text''': Value('''string''' ), '''title''': Value('''string''' ), '''embeddings''': Sequence(Value('''float32''' ) )} ) # optional, save as float32 instead of float64 to save space _a = dataset.map( partial(UpperCamelCase , ctx_encoder=UpperCamelCase , ctx_tokenizer=UpperCamelCase ) , batched=UpperCamelCase , batch_size=processing_args.batch_size , features=UpperCamelCase , ) # And finally save your dataset _a = os.path.join(rag_example_args.output_dir , '''my_knowledge_dataset''' ) dataset.save_to_disk(UpperCamelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('''Step 2 - Index the dataset''' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _a = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('''embeddings''' , custom_index=UpperCamelCase ) # And save the index _a = os.path.join(rag_example_args.output_dir , '''my_knowledge_dataset_hnsw_index.faiss''' ) dataset.get_index('''embeddings''' ).save(UpperCamelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A : lowercase_ = field( default=str(Path(_a ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) ,metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} ,) lowercase_ = field( default=_a ,metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} ,) lowercase_ = field( default='facebook/rag-sequence-nq' ,metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} ,) lowercase_ = field( default='facebook/dpr-ctx_encoder-multiset-base' ,metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } ,) lowercase_ = field( default=str(Path(_a ).parent / 'test_run' / 'dummy-kb' ) ,metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} ,) @dataclass class A : lowercase_ = field( default=_a ,metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } ,) lowercase_ = field( default=16 ,metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } ,) @dataclass class A : lowercase_ = field( default=768 ,metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} ,) lowercase_ = field( default=128 ,metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } ,) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) _snake_case : Optional[int] = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) _snake_case : Optional[int] = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: _snake_case : List[Any] = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
351
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _snake_case : Dict = logging.get_logger(__name__) _snake_case : Optional[Any] = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class A ( _a ,_a ): lowercase_ = 'nat' lowercase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[str] , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any=3 , lowerCAmelCase_ : List[Any]=64 , lowerCAmelCase_ : Dict=[3, 4, 6, 5] , lowerCAmelCase_ : Dict=[2, 4, 8, 16] , lowerCAmelCase_ : str=7 , lowerCAmelCase_ : Dict=3.0 , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Tuple=0.0 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : int="gelu" , lowerCAmelCase_ : List[str]=0.0_2 , lowerCAmelCase_ : str=1e-5 , lowerCAmelCase_ : Tuple=0.0 , lowerCAmelCase_ : str=None , lowerCAmelCase_ : int=None , **lowerCAmelCase_ : List[Any] , ) -> Any: """simple docstring""" super().__init__(**lowerCAmelCase_ ) _a = patch_size _a = num_channels _a = embed_dim _a = depths _a = len(lowerCAmelCase_ ) _a = num_heads _a = kernel_size _a = mlp_ratio _a = qkv_bias _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = drop_path_rate _a = hidden_act _a = layer_norm_eps _a = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _a = int(embed_dim * 2 ** (len(lowerCAmelCase_ ) - 1) ) _a = layer_scale_init_value _a = ['''stem'''] + [F'stage{idx}' for idx in range(1 , len(lowerCAmelCase_ ) + 1 )] _a , _a = get_aligned_output_features_output_indices( out_features=lowerCAmelCase_ , out_indices=lowerCAmelCase_ , stage_names=self.stage_names )
179
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __snake_case =get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class UpperCAmelCase_ ( __lowercase , unittest.TestCase ): lowerCamelCase : Union[str, Any] = XLMRobertaTokenizer lowerCamelCase : Dict = XLMRobertaTokenizerFast lowerCamelCase : str = True lowerCamelCase : Optional[int] = True def __UpperCAmelCase ( self : Dict ) -> str: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase = XLMRobertaTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCAmelCase ( self : List[Any] ) -> Union[str, Any]: lowerCAmelCase = '<pad>' lowerCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_0_0_2 ) def __UpperCAmelCase ( self : Union[str, Any] ) -> Tuple: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def __UpperCAmelCase ( self : Optional[int] ) -> List[str]: lowerCAmelCase = XLMRobertaTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) lowerCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def __UpperCAmelCase ( self : List[Any] ) -> str: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-xlm-roberta', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) lowerCAmelCase = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) lowerCAmelCase = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(UpperCAmelCase__ , UpperCAmelCase__ ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCAmelCase__ ) # Save tokenizer rust, legacy_format=True lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(UpperCAmelCase__ , legacy_format=UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it save with the same files self.assertSequenceEqual(UpperCAmelCase__ , UpperCAmelCase__ ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) shutil.rmtree(UpperCAmelCase__ ) # Save tokenizer rust, legacy_format=False lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(UpperCAmelCase__ , legacy_format=UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(UpperCAmelCase__ ) lowerCAmelCase = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) shutil.rmtree(UpperCAmelCase__ ) @cached_property def __UpperCAmelCase ( self : Union[str, Any] ) -> Dict: return XLMRobertaTokenizer.from_pretrained('xlm-roberta-base' ) def __UpperCAmelCase ( self : int ) -> Union[str, Any]: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(UpperCAmelCase__ , f.name ) lowerCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=UpperCAmelCase__ ) lowerCAmelCase = pickle.dumps(UpperCAmelCase__ ) pickle.loads(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[int] ) -> Dict: if not self.test_rust_tokenizer: return lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = 'I was born in 92000, and this is falsé.' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase__ ) lowerCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) lowerCAmelCase = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = tokenizer.encode(UpperCAmelCase__ ) lowerCAmelCase = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def __UpperCAmelCase ( self : Optional[Any] ) -> Any: lowerCAmelCase = 'Hello World!' lowerCAmelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def __UpperCAmelCase ( self : Tuple ) -> Optional[int]: lowerCAmelCase = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCAmelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def __UpperCAmelCase ( self : str ) -> Tuple: # fmt: off lowerCAmelCase = {'input_ids': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='xlm-roberta-base' , revision='d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3' , )
4
_A = [0, 2, 4, 6, 8] _A = [1, 3, 5, 7, 9] def lowerCamelCase__ ( a__ : int , a__ : int , a__ : list[int] , a__ : int ) -> int: if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 UpperCamelCase_ = 0 for digit in range(10 ): UpperCamelCase_ = digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , a__ , a__ ) return result UpperCamelCase_ = 0 for digita in range(10 ): UpperCamelCase_ = digita if (remainder + digita) % 2 == 0: UpperCamelCase_ = ODD_DIGITS else: UpperCamelCase_ = EVEN_DIGITS for digita in other_parity_digits: UpperCamelCase_ = digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , a__ , a__ , ) return result def lowerCamelCase__ ( a__ : int = 9 ) -> int: UpperCamelCase_ = 0 for length in range(1 , max_power + 1 ): result += reversible_numbers(a__ , 0 , [0] * length , a__ ) return result if __name__ == "__main__": print(F'''{solution() = }''')
122
0
'''simple docstring''' from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function _SCREAMING_SNAKE_CASE : Optional[Any] = 1.054571817e-34 # unit of ℏ : J * s _SCREAMING_SNAKE_CASE : int = 3e8 # unit of c : m * s^-1 def UpperCamelCase_( snake_case : float , snake_case : float , snake_case : float ): '''simple docstring''' if (force, area, distance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if force < 0: raise ValueError("Magnitude of force can not be negative" ) if distance < 0: raise ValueError("Distance can not be negative" ) if area < 0: raise ValueError("Area can not be negative" ) if force == 0: snake_case_ = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 2_4_0 * (distance) ** 4 ) return {"force": force} elif area == 0: snake_case_ = (2_4_0 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: snake_case_ = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (2_4_0 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError("One and only one argument must be 0" ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
367
'''simple docstring''' from random import shuffle import tensorflow as tf from numpy import array def UpperCamelCase_( snake_case : Optional[int] , snake_case : Optional[int] ): '''simple docstring''' snake_case_ = int(snake_case ) assert noofclusters < len(snake_case ) # Find out the dimensionality snake_case_ = len(vectors[0] ) # Will help select random centroids from among the available vectors snake_case_ = list(range(len(snake_case ) ) ) shuffle(snake_case ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. snake_case_ = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION snake_case_ = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points snake_case_ = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(snake_case ) ] ##These nodes will assign the centroid Variables the appropriate ##values snake_case_ = tf.placeholder("float64" , [dim] ) snake_case_ = [] for centroid in centroids: cent_assigns.append(tf.assign(snake_case , snake_case ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) snake_case_ = [tf.Variable(0 ) for i in range(len(snake_case ) )] ##These nodes will assign an assignment Variable the appropriate ##value snake_case_ = tf.placeholder("int32" ) snake_case_ = [] for assignment in assignments: cluster_assigns.append(tf.assign(snake_case , snake_case ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input snake_case_ = tf.placeholder("float" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors snake_case_ = tf.reduce_mean(snake_case , 0 ) ##Node for computing Euclidean distances # Placeholders for input snake_case_ = tf.placeholder("float" , [dim] ) snake_case_ = tf.placeholder("float" , [dim] ) snake_case_ = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(snake_case , snake_case ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input snake_case_ = tf.placeholder("float" , [noofclusters] ) snake_case_ = tf.argmin(snake_case , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. snake_case_ = tf.initialize_all_variables() # Initialize all variables sess.run(snake_case ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. snake_case_ = 1_0_0 for _ in range(snake_case ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(snake_case ) ): snake_case_ = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. snake_case_ = [ sess.run(snake_case , feed_dict={va: vect, va: sess.run(snake_case )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input snake_case_ = sess.run( snake_case , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(snake_case ): # Collect all the vectors assigned to this cluster snake_case_ = [ vectors[i] for i in range(len(snake_case ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location snake_case_ = sess.run( snake_case , feed_dict={mean_input: array(snake_case )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments snake_case_ = sess.run(snake_case ) snake_case_ = sess.run(snake_case ) return centroids, assignments
92
0
"""simple docstring""" import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename _a = """http://www.mocksite.com/file1.txt""" _a = """\"text\": [\"foo\", \"foo\"]""" _a = """6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8""" class _UpperCAmelCase: lowercase__ = 2_00 lowercase__ = {'Content-Length': '100'} lowercase__ = {} def UpperCAmelCase ( self , **__a) -> Optional[int]: '''simple docstring''' return [bytes(__a , '''utf-8''')] def lowerCamelCase__ ( *__snake_case, **__snake_case ) -> int: """simple docstring""" return MockResponse() @pytest.mark.parametrize('''urls_type''', [str, list, dict] ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Dict: """simple docstring""" import requests monkeypatch.setattr(__snake_case, '''request''', __snake_case ) _UpperCamelCase = URL if issubclass(__snake_case, __snake_case ): _UpperCamelCase = url elif issubclass(__snake_case, __snake_case ): _UpperCamelCase = [url] elif issubclass(__snake_case, __snake_case ): _UpperCamelCase = {'''train''': url} _UpperCamelCase = '''dummy''' _UpperCamelCase = '''downloads''' _UpperCamelCase = tmp_path _UpperCamelCase = DownloadConfig( cache_dir=os.path.join(__snake_case, __snake_case ), use_etag=__snake_case, ) _UpperCamelCase = DownloadManager(dataset_name=__snake_case, download_config=__snake_case ) _UpperCamelCase = dl_manager.download(__snake_case ) _UpperCamelCase = urls for downloaded_paths in [downloaded_paths]: if isinstance(__snake_case, __snake_case ): _UpperCamelCase = [downloaded_paths] _UpperCamelCase = [urls] elif isinstance(__snake_case, __snake_case ): assert "train" in downloaded_paths.keys() _UpperCamelCase = downloaded_paths.values() _UpperCamelCase = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(__snake_case, __snake_case ): assert downloaded_path == dl_manager.downloaded_paths[input_url] _UpperCamelCase = Path(__snake_case ) _UpperCamelCase = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() _UpperCamelCase = downloaded_path.read_text() assert content == CONTENT _UpperCamelCase = downloaded_path.with_suffix('''.json''' ) assert metadata_downloaded_path.exists() _UpperCamelCase = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('''paths_type''', [str, list, dict] ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> List[str]: """simple docstring""" _UpperCamelCase = str(__snake_case ) if issubclass(__snake_case, __snake_case ): _UpperCamelCase = filename elif issubclass(__snake_case, __snake_case ): _UpperCamelCase = [filename] elif issubclass(__snake_case, __snake_case ): _UpperCamelCase = {'''train''': filename} _UpperCamelCase = '''dummy''' _UpperCamelCase = xz_file.parent _UpperCamelCase = '''extracted''' _UpperCamelCase = DownloadConfig( cache_dir=__snake_case, use_etag=__snake_case, ) _UpperCamelCase = DownloadManager(dataset_name=__snake_case, download_config=__snake_case ) _UpperCamelCase = dl_manager.extract(__snake_case ) _UpperCamelCase = paths for extracted_paths in [extracted_paths]: if isinstance(__snake_case, __snake_case ): _UpperCamelCase = [extracted_paths] _UpperCamelCase = [paths] elif isinstance(__snake_case, __snake_case ): assert "train" in extracted_paths.keys() _UpperCamelCase = extracted_paths.values() _UpperCamelCase = paths.values() assert extracted_paths for extracted_path, input_path in zip(__snake_case, __snake_case ): assert extracted_path == dl_manager.extracted_paths[input_path] _UpperCamelCase = Path(__snake_case ) _UpperCamelCase = extracted_path.parts assert parts[-1] == hash_url_to_filename(__snake_case, etag=__snake_case ) assert parts[-2] == extracted_subdir assert extracted_path.exists() _UpperCamelCase = extracted_path.read_text() _UpperCamelCase = text_file.read_text() assert extracted_file_content == expected_file_content def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert path.endswith('''.jsonl''' ) for num_items, line in enumerate(__snake_case, start=1 ): _UpperCamelCase = json.loads(line.decode('''utf-8''' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('''archive_jsonl''', ['''tar_jsonl_path''', '''zip_jsonl_path'''] ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = request.getfixturevalue(__snake_case ) _UpperCamelCase = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(__snake_case ), start=1 ): _test_jsonl(__snake_case, __snake_case ) assert num_jsonl == 2 @pytest.mark.parametrize('''archive_nested_jsonl''', ['''tar_nested_jsonl_path''', '''zip_nested_jsonl_path'''] ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = request.getfixturevalue(__snake_case ) _UpperCamelCase = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(__snake_case ), start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(__snake_case ), start=1 ): _test_jsonl(__snake_case, __snake_case ) assert num_tar == 1 assert num_jsonl == 2 def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(__snake_case ), start=1 ): assert os.path.basename(__snake_case ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
194
"""simple docstring""" import datasets _a = """\ @InProceedings{conneau2018xnli, author = \"Conneau, Alexis and Rinott, Ruty and Lample, Guillaume and Williams, Adina and Bowman, Samuel R. and Schwenk, Holger and Stoyanov, Veselin\", title = \"XNLI: Evaluating Cross-lingual Sentence Representations\", booktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing\", year = \"2018\", publisher = \"Association for Computational Linguistics\", location = \"Brussels, Belgium\", } """ _a = """\ XNLI is a subset of a few thousand examples from MNLI which has been translated into a 14 different languages (some low-ish resource). As with MNLI, the goal is to predict textual entailment (does sentence A imply/contradict/neither sentence B) and is a classification task (given two sentences, predict one of three labels). """ _a = """ Computes XNLI score which is just simple accuracy. Args: predictions: Predicted labels. references: Ground truth labels. Returns: 'accuracy': accuracy Examples: >>> predictions = [0, 1] >>> references = [0, 1] >>> xnli_metric = datasets.load_metric(\"xnli\") >>> results = xnli_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} """ def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase( datasets.Metric ): def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32'''), '''references''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32'''), }) , codebase_urls=[] , reference_urls=[] , format='''numpy''' , ) def UpperCAmelCase ( self , __a , __a) -> Dict: '''simple docstring''' return {"accuracy": simple_accuracy(__a , __a)}
194
1
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = 0 SCREAMING_SNAKE_CASE_ : Dict = 1 SCREAMING_SNAKE_CASE_ : Union[str, Any] = 2 @add_end_docstrings(lowerCamelCase_ ) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Dict = """ In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas's young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> """ def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == 'tf' else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. SCREAMING_SNAKE_CASE = None if self.model.config.prefix is not None: SCREAMING_SNAKE_CASE = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. SCREAMING_SNAKE_CASE = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._sanitize_parameters(prefix=lowerCAmelCase__ , **self._forward_params ) SCREAMING_SNAKE_CASE = {**self._preprocess_params, **preprocess_params} SCREAMING_SNAKE_CASE = {**self._forward_params, **forward_params} def __A ( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> Optional[Any]: SCREAMING_SNAKE_CASE = {} if prefix is not None: SCREAMING_SNAKE_CASE = prefix if prefix: SCREAMING_SNAKE_CASE = self.tokenizer( lowerCAmelCase__ , padding=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=self.framework ) SCREAMING_SNAKE_CASE = prefix_inputs['input_ids'].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( F'{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected' ' [None, \'hole\']' ) SCREAMING_SNAKE_CASE = handle_long_generation preprocess_params.update(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = generate_kwargs SCREAMING_SNAKE_CASE = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError('`return_text` is mutually exclusive with `return_full_text`' ) if return_tensors is not None: raise ValueError('`return_full_text` is mutually exclusive with `return_tensors`' ) SCREAMING_SNAKE_CASE = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError('`return_text` is mutually exclusive with `return_tensors`' ) SCREAMING_SNAKE_CASE = ReturnType.TENSORS if return_type is not None: SCREAMING_SNAKE_CASE = return_type if clean_up_tokenization_spaces is not None: SCREAMING_SNAKE_CASE = clean_up_tokenization_spaces if stop_sequence is not None: SCREAMING_SNAKE_CASE = self.tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > 1: warnings.warn( 'Stopping on a multiple token sequence is not yet supported on transformers. The first token of' ' the stop sequence will be used as the stop sequence string in the interim.' ) SCREAMING_SNAKE_CASE = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __A ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[int]: # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({'add_space_before_punct_symbol': True} ) return super()._parse_and_tokenize(*lowerCAmelCase__ , **lowerCAmelCase__ ) def __call__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: return super().__call__(lowerCAmelCase__ , **lowerCAmelCase__ ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__="" , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Dict: SCREAMING_SNAKE_CASE = self.tokenizer( prefix + prompt_text , padding=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=self.framework ) SCREAMING_SNAKE_CASE = prompt_text if handle_long_generation == "hole": SCREAMING_SNAKE_CASE = inputs['input_ids'].shape[-1] if "max_new_tokens" in generate_kwargs: SCREAMING_SNAKE_CASE = generate_kwargs['max_new_tokens'] else: SCREAMING_SNAKE_CASE = generate_kwargs.get('max_length' , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError('We cannot infer how many new tokens are expected' ) if cur_len + new_tokens > self.tokenizer.model_max_length: SCREAMING_SNAKE_CASE = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( 'We cannot use `hole` to handle this generation the number of desired tokens exceeds the' ' models max length' ) SCREAMING_SNAKE_CASE = inputs['input_ids'][:, -keep_length:] if "attention_mask" in inputs: SCREAMING_SNAKE_CASE = inputs['attention_mask'][:, -keep_length:] return inputs def __A ( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> str: SCREAMING_SNAKE_CASE = model_inputs['input_ids'] SCREAMING_SNAKE_CASE = model_inputs.get('attention_mask' , lowerCAmelCase__ ) # Allow empty prompts if input_ids.shape[1] == 0: SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = 1 else: SCREAMING_SNAKE_CASE = input_ids.shape[0] SCREAMING_SNAKE_CASE = model_inputs.pop('prompt_text' ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. SCREAMING_SNAKE_CASE = generate_kwargs.pop('prefix_length' , 0 ) if prefix_length > 0: SCREAMING_SNAKE_CASE = 'max_new_tokens' in generate_kwargs or ( 'generation_config' in generate_kwargs and generate_kwargs['generation_config'].max_new_tokens is not None ) if not has_max_new_tokens: SCREAMING_SNAKE_CASE = generate_kwargs.get('max_length' ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length SCREAMING_SNAKE_CASE = 'min_new_tokens' in generate_kwargs or ( 'generation_config' in generate_kwargs and generate_kwargs['generation_config'].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL SCREAMING_SNAKE_CASE = self.model.generate(input_ids=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , **lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = generated_sequence.shape[0] if self.framework == "pt": SCREAMING_SNAKE_CASE = generated_sequence.reshape(lowerCAmelCase__ , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": SCREAMING_SNAKE_CASE = tf.reshape(lowerCAmelCase__ , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def __A ( self , lowerCAmelCase__ , lowerCAmelCase__=ReturnType.FULL_TEXT , lowerCAmelCase__=True ) -> Optional[Any]: SCREAMING_SNAKE_CASE = model_outputs['generated_sequence'][0] SCREAMING_SNAKE_CASE = model_outputs['input_ids'] SCREAMING_SNAKE_CASE = model_outputs['prompt_text'] SCREAMING_SNAKE_CASE = generated_sequence.numpy().tolist() SCREAMING_SNAKE_CASE = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: SCREAMING_SNAKE_CASE = {'generated_token_ids': sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text SCREAMING_SNAKE_CASE = self.tokenizer.decode( lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: SCREAMING_SNAKE_CASE = 0 else: SCREAMING_SNAKE_CASE = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ , ) ) if return_type == ReturnType.FULL_TEXT: SCREAMING_SNAKE_CASE = prompt_text + text[prompt_length:] else: SCREAMING_SNAKE_CASE = text[prompt_length:] SCREAMING_SNAKE_CASE = {'generated_text': all_text} records.append(lowerCAmelCase__ ) return records
38
"""simple docstring""" from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> bool: SCREAMING_SNAKE_CASE = int(number**0.5 ) return number == sq * sq def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> tuple[int, int]: SCREAMING_SNAKE_CASE = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den SCREAMING_SNAKE_CASE = x_den * y_den * z_den SCREAMING_SNAKE_CASE = gcd(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) top //= hcf bottom //= hcf return top, bottom def lowercase (SCREAMING_SNAKE_CASE_ : int = 35 ) -> int: SCREAMING_SNAKE_CASE = set() SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = Fraction(0 ) SCREAMING_SNAKE_CASE = 42 for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 SCREAMING_SNAKE_CASE = x_num * y_den + x_den * y_num SCREAMING_SNAKE_CASE = x_den * y_den SCREAMING_SNAKE_CASE = gcd(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: SCREAMING_SNAKE_CASE = add_three( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) unique_s.add(SCREAMING_SNAKE_CASE_ ) # n=2 SCREAMING_SNAKE_CASE = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) SCREAMING_SNAKE_CASE = x_den * x_den * y_den * y_den if is_sq(SCREAMING_SNAKE_CASE_ ) and is_sq(SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = int(sqrt(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = int(sqrt(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = gcd(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: SCREAMING_SNAKE_CASE = add_three( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) unique_s.add(SCREAMING_SNAKE_CASE_ ) # n=-1 SCREAMING_SNAKE_CASE = x_num * y_num SCREAMING_SNAKE_CASE = x_den * y_num + x_num * y_den SCREAMING_SNAKE_CASE = gcd(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: SCREAMING_SNAKE_CASE = add_three( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) unique_s.add(SCREAMING_SNAKE_CASE_ ) # n=2 SCREAMING_SNAKE_CASE = x_num * x_num * y_num * y_num SCREAMING_SNAKE_CASE = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(SCREAMING_SNAKE_CASE_ ) and is_sq(SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = int(sqrt(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = int(sqrt(SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE = gcd(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: SCREAMING_SNAKE_CASE = add_three( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) unique_s.add(SCREAMING_SNAKE_CASE_ ) for num, den in unique_s: total += Fraction(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return total.denominator + total.numerator if __name__ == "__main__": print(f'''{solution() = }''')
38
1
'''simple docstring''' def _A ( snake_case ) -> int: _lowercase : Tuple = abs(snake_case ) _lowercase : int = 0 while n > 0: res += n % 10 n //= 10 return res def _A ( snake_case ) -> Any: _lowercase : Dict = abs(snake_case ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def _A ( snake_case ) -> Optional[int]: return sum(int(snake_case ) for c in str(abs(snake_case ) ) ) def _A ( ) -> Union[str, Any]: from collections.abc import Callable from timeit import timeit def benchmark_a_function(snake_case , snake_case ) -> None: _lowercase : int = F'''{func.__name__}({value})''' _lowercase : str = timeit(F'''__main__.{call}''' , setup="import __main__" ) print(F'''{call:56} = {func(snake_case )} -- {timing:.4f} seconds''' ) for value in (26_21_44, 11_25_89_99_06_84_26_24, 1_26_76_50_60_02_28_22_94_01_49_67_03_20_53_76): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(snake_case , snake_case ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
250
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCamelCase : Optional[Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = ['''pixel_values'''] def __init__( self : List[Any] , A_ : bool = True , A_ : Dict[str, int] = None , A_ : PILImageResampling = PILImageResampling.BICUBIC , A_ : bool = True , A_ : Dict[str, int] = None , A_ : bool = True , A_ : Union[int, float] = 1 / 255 , A_ : bool = True , A_ : Optional[Union[float, List[float]]] = None , A_ : Optional[Union[float, List[float]]] = None , A_ : bool = True , **A_ : Dict , ) -> None: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = size if size is not None else {'shortest_edge': 224} lowerCamelCase_ = get_size_dict(A_ , default_to_square=A_ ) lowerCamelCase_ = crop_size if crop_size is not None else {'height': 224, 'width': 224} lowerCamelCase_ = get_size_dict(A_ , default_to_square=A_ , param_name='crop_size' ) lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = resample lowerCamelCase_ = do_center_crop lowerCamelCase_ = crop_size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowerCamelCase_ = image_std if image_std is not None else OPENAI_CLIP_STD lowerCamelCase_ = do_convert_rgb def a__ ( self : Dict , A_ : np.ndarray , A_ : Dict[str, int] , A_ : PILImageResampling = PILImageResampling.BICUBIC , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : Union[str, Any] , ) -> np.ndarray: """simple docstring""" lowerCamelCase_ = get_size_dict(A_ , default_to_square=A_ ) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) lowerCamelCase_ = get_resize_output_image_size(A_ , size=size['shortest_edge'] , default_to_square=A_ ) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def a__ ( self : Tuple , A_ : np.ndarray , A_ : Dict[str, int] , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : Dict , ) -> np.ndarray: """simple docstring""" lowerCamelCase_ = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(A_ , size=(size['height'], size['width']) , data_format=A_ , **A_ ) def a__ ( self : str , A_ : np.ndarray , A_ : Union[int, float] , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : Union[str, Any] , ) -> str: """simple docstring""" return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def a__ ( self : Optional[Any] , A_ : np.ndarray , A_ : Union[float, List[float]] , A_ : Union[float, List[float]] , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : str , ) -> np.ndarray: """simple docstring""" return normalize(A_ , mean=A_ , std=A_ , data_format=A_ , **A_ ) def a__ ( self : Any , A_ : ImageInput , A_ : bool = None , A_ : Dict[str, int] = None , A_ : PILImageResampling = None , A_ : bool = None , A_ : int = None , A_ : bool = None , A_ : float = None , A_ : bool = None , A_ : Optional[Union[float, List[float]]] = None , A_ : Optional[Union[float, List[float]]] = None , A_ : bool = None , A_ : Optional[Union[str, TensorType]] = None , A_ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A_ : Dict , ) -> PIL.Image.Image: """simple docstring""" lowerCamelCase_ = do_resize if do_resize is not None else self.do_resize lowerCamelCase_ = size if size is not None else self.size lowerCamelCase_ = get_size_dict(A_ , param_name='size' , default_to_square=A_ ) lowerCamelCase_ = resample if resample is not None else self.resample lowerCamelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCamelCase_ = crop_size if crop_size is not None else self.crop_size lowerCamelCase_ = get_size_dict(A_ , param_name='crop_size' , default_to_square=A_ ) lowerCamelCase_ = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase_ = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase_ = image_mean if image_mean is not None else self.image_mean lowerCamelCase_ = image_std if image_std is not None else self.image_std lowerCamelCase_ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCamelCase_ = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCamelCase_ = [convert_to_rgb(A_ ) for image in images] # All transformations expect numpy arrays. lowerCamelCase_ = [to_numpy_array(A_ ) for image in images] if do_resize: lowerCamelCase_ = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] if do_center_crop: lowerCamelCase_ = [self.center_crop(image=A_ , size=A_ ) for image in images] if do_rescale: lowerCamelCase_ = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_normalize: lowerCamelCase_ = [self.normalize(image=A_ , mean=A_ , std=A_ ) for image in images] lowerCamelCase_ = [to_channel_dimension_format(A_ , A_ ) for image in images] lowerCamelCase_ = {'pixel_values': images} return BatchFeature(data=A_ , tensor_type=A_ )
204
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase : Dict = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[int] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _lowercase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
21
'''simple docstring''' import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class __magic_name__ ( unittest.TestCase): @parameterized.expand([(None,), ("""foo.json""",)] ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] , lowercase_ : str ): lowercase_ : Union[str, Any] = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowercase_ , config_name=lowercase_ ) lowercase_ : Optional[int] = GenerationConfig.from_pretrained(lowercase_ , config_name=lowercase_ ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , lowercase_ ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 50 ) self.assertEqual(loaded_config.max_length , 20 ) self.assertEqual(loaded_config.max_time , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : int = AutoConfig.from_pretrained("""gpt2""" ) lowercase_ : List[Any] = GenerationConfig.from_model_config(lowercase_ ) lowercase_ : Optional[int] = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(lowercase_ , lowercase_ ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def SCREAMING_SNAKE_CASE_ ( self : List[str] ): lowercase_ : Optional[int] = GenerationConfig() lowercase_ : int = { """max_new_tokens""": 1024, """foo""": """bar""", } lowercase_ : List[str] = copy.deepcopy(lowercase_ ) lowercase_ : Tuple = generation_config.update(**lowercase_ ) # update_kwargs was not modified (no side effects) self.assertEqual(lowercase_ , lowercase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 1024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(lowercase_ , {"""foo""": """bar"""} ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Dict = GenerationConfig() lowercase_ : int = """bar""" with tempfile.TemporaryDirectory("""test-generation-config""" ) as tmp_dir: generation_config.save_pretrained(lowercase_ ) lowercase_ : Optional[int] = GenerationConfig.from_pretrained(lowercase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , """bar""" ) lowercase_ : List[str] = GenerationConfig.from_model_config(lowercase_ ) assert not hasattr(lowercase_ , """foo""" ) # no new kwargs should be initialized if from config def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Optional[int] = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , lowercase_ ) self.assertEqual(default_config.num_beams , 1 ) lowercase_ : Dict = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , lowercase_ ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowercase_ ) lowercase_ : Tuple = GenerationConfig.from_pretrained(lowercase_ , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , lowercase_ ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class __magic_name__ ( unittest.TestCase): @classmethod def SCREAMING_SNAKE_CASE_ ( cls : Any ): lowercase_ : int = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def SCREAMING_SNAKE_CASE_ ( cls : List[Any] ): try: delete_repo(token=cls._token , repo_id="""test-generation-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-generation-config-org""" ) except HTTPError: pass def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : Tuple = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""test-generation-config""" , use_auth_token=self._token ) lowercase_ : List[Any] = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-generation-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowercase_ , repo_id="""test-generation-config""" , push_to_hub=lowercase_ , use_auth_token=self._token ) lowercase_ : int = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : List[Any] = GenerationConfig( do_sample=lowercase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""valid_org/test-generation-config-org""" , use_auth_token=self._token ) lowercase_ : Optional[Any] = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-generation-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowercase_ , repo_id="""valid_org/test-generation-config-org""" , push_to_hub=lowercase_ , use_auth_token=self._token ) lowercase_ : int = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowercase_ , getattr(lowercase_ , lowercase_ ) )
21
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __snake_case ( _lowerCamelCase ): __lowerCamelCase = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
143
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Tuple = logging.get_logger(__name__) lowerCAmelCase__ : Union[str, Any] = {'''vocab_file''': '''sentencepiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ : Dict = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, '''tokenizer_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ : List[Any] = { '''google/rembert''': 2_56, } lowerCAmelCase__ : List[str] = '''▁''' class __snake_case ( _lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RemBertTokenizer def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase="[CLS]" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<unk>" , __UpperCamelCase="[SEP]" , __UpperCamelCase="<pad>" , __UpperCamelCase="[CLS]" , __UpperCamelCase="[MASK]" , **__UpperCamelCase , ) -> Tuple: '''simple docstring''' snake_case__ : Optional[Any] = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token super().__init__( __UpperCamelCase , tokenizer_file=__UpperCamelCase , do_lower_case=__UpperCamelCase , remove_space=__UpperCamelCase , keep_accents=__UpperCamelCase , bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , pad_token=__UpperCamelCase , cls_token=__UpperCamelCase , mask_token=__UpperCamelCase , **__UpperCamelCase , ) snake_case__ : int = do_lower_case snake_case__ : Any = remove_space snake_case__ : List[Any] = keep_accents snake_case__ : Dict = vocab_file snake_case__ : int = False if not self.vocab_file else True def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : Dict = [self.sep_token_id] snake_case__ : List[str] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__UpperCamelCase )) + [1] + ([0] * len(__UpperCamelCase )) + [1] return [1] + ([0] * len(__UpperCamelCase )) + [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> List[int]: '''simple docstring''' snake_case__ : List[Any] = [self.sep_token_id] snake_case__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(__UpperCamelCase ): logger.error('Vocabulary path ({}) should be a directory'.format(__UpperCamelCase ) ) return snake_case__ : List[str] = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ): copyfile(self.vocab_file , __UpperCamelCase ) return (out_vocab_file,)
143
1
def SCREAMING_SNAKE_CASE ( snake_case_ : int , snake_case_ : float , snake_case_ : float ): return round(float(moles / volume ) * nfactor ) def SCREAMING_SNAKE_CASE ( snake_case_ : float , snake_case_ : float , snake_case_ : float ): return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def SCREAMING_SNAKE_CASE ( snake_case_ : float , snake_case_ : float , snake_case_ : float ): return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def SCREAMING_SNAKE_CASE ( snake_case_ : float , snake_case_ : float , snake_case_ : float ): return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
286
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Optional[int] = logging.get_logger(__name__) __lowerCamelCase : str = { """andreasmadsen/efficient_mlm_m0.40""": ( """https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json""" ), } class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ ): """simple docstring""" a_ = "roberta-prelayernorm" def __init__( self : Tuple , __A : Any=5_0_2_6_5 , __A : Optional[int]=7_6_8 , __A : Dict=1_2 , __A : Union[str, Any]=1_2 , __A : List[Any]=3_0_7_2 , __A : Optional[Any]="gelu" , __A : Optional[int]=0.1 , __A : Tuple=0.1 , __A : Optional[Any]=5_1_2 , __A : List[str]=2 , __A : Optional[int]=0.0_2 , __A : Tuple=1e-1_2 , __A : Any=1 , __A : str=0 , __A : int=2 , __A : List[str]="absolute" , __A : Optional[Any]=True , __A : List[Any]=None , **__A : Optional[Any] , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) snake_case__ : Tuple = vocab_size snake_case__ : Optional[Any] = hidden_size snake_case__ : List[Any] = num_hidden_layers snake_case__ : List[Any] = num_attention_heads snake_case__ : Dict = hidden_act snake_case__ : Union[str, Any] = intermediate_size snake_case__ : List[Any] = hidden_dropout_prob snake_case__ : Any = attention_probs_dropout_prob snake_case__ : int = max_position_embeddings snake_case__ : Tuple = type_vocab_size snake_case__ : Optional[int] = initializer_range snake_case__ : int = layer_norm_eps snake_case__ : Dict = position_embedding_type snake_case__ : int = use_cache snake_case__ : Dict = classifier_dropout class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ ): """simple docstring""" @property def _lowercase ( self : Optional[int] ): if self.task == "multiple-choice": snake_case__ : List[str] = {0: "batch", 1: "choice", 2: "sequence"} else: snake_case__ : Tuple = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
286
1
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch __snake_case = random.Random() def a ( __a , __a=1.0 , __a=None , __a=None ) -> Any: '''simple docstring''' if rng is None: UpperCamelCase__ :Any = global_rng UpperCamelCase__ :Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self , UpperCamelCase_ , UpperCamelCase_=7 , UpperCamelCase_=400 , UpperCamelCase_=2000 , UpperCamelCase_=1 , UpperCamelCase_=0.0 , UpperCamelCase_=16000 , UpperCamelCase_=True , UpperCamelCase_=80 , UpperCamelCase_=16 , UpperCamelCase_=64 , UpperCamelCase_="hann_window" , UpperCamelCase_=80 , UpperCamelCase_=7600 , UpperCamelCase_=1e-10 , UpperCamelCase_=True , ): '''simple docstring''' UpperCamelCase__ :str = parent UpperCamelCase__ :List[Any] = batch_size UpperCamelCase__ :str = min_seq_length UpperCamelCase__ :Optional[Any] = max_seq_length UpperCamelCase__ :Dict = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCamelCase__ :List[Any] = feature_size UpperCamelCase__ :Union[str, Any] = padding_value UpperCamelCase__ :Any = sampling_rate UpperCamelCase__ :Tuple = do_normalize UpperCamelCase__ :int = num_mel_bins UpperCamelCase__ :Tuple = hop_length UpperCamelCase__ :Any = win_length UpperCamelCase__ :int = win_function UpperCamelCase__ :Optional[Any] = fmin UpperCamelCase__ :List[str] = fmax UpperCamelCase__ :Tuple = mel_floor UpperCamelCase__ :Dict = return_attention_mask def lowerCAmelCase__ ( self ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def lowerCAmelCase__ ( self , UpperCamelCase_=False , UpperCamelCase_=False ): '''simple docstring''' def _flatten(UpperCamelCase_ ): return list(itertools.chain(*UpperCamelCase_ ) ) if equal_length: UpperCamelCase__ :Optional[int] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size UpperCamelCase__ :List[str] = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase__ :int = [np.asarray(UpperCamelCase_ ) for x in speech_inputs] return speech_inputs def lowerCAmelCase__ ( self , UpperCamelCase_=False , UpperCamelCase_=False ): '''simple docstring''' if equal_length: UpperCamelCase__ :int = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size UpperCamelCase__ :Dict = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase__ :Optional[Any] = [np.asarray(UpperCamelCase_ ) for x in speech_inputs] return speech_inputs @require_torch class lowercase ( _a , unittest.TestCase ): """simple docstring""" _a = SpeechTaFeatureExtractor def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = SpeechTaFeatureExtractionTester(self ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' self.assertTrue(np.all(np.mean(UpperCamelCase_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(UpperCamelCase_ , axis=0 ) - 1 ) < 1e-3 ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase__ :Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :Any = [np.asarray(UpperCamelCase_ ) for speech_input in speech_inputs] # Test not batched input UpperCamelCase__ :Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values UpperCamelCase__ :Union[str, Any] = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3 ) ) # Test batched UpperCamelCase__ :int = feat_extract(UpperCamelCase_ , return_tensors='''np''' ).input_values UpperCamelCase__ :Tuple = feat_extract(UpperCamelCase_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_ ): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3 ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ :Union[str, Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :str = ['longest', 'max_length', 'do_not_pad'] UpperCamelCase__ :List[str] = [None, 1600, None] for max_length, padding in zip(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Any = feat_extract(UpperCamelCase_ , padding=UpperCamelCase_ , max_length=UpperCamelCase_ , return_tensors='''np''' ) UpperCamelCase__ :Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ :Tuple = range(800 , 1400 , 200 ) UpperCamelCase__ :Optional[int] = [floats_list((1, x) )[0] for x in lengths] UpperCamelCase__ :Optional[Any] = ['longest', 'max_length', 'do_not_pad'] UpperCamelCase__ :Optional[int] = [None, 1600, None] for max_length, padding in zip(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Union[str, Any] = feat_extract(UpperCamelCase_ , max_length=UpperCamelCase_ , padding=UpperCamelCase_ ) UpperCamelCase__ :List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ :Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :Union[str, Any] = feat_extract( UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) UpperCamelCase__ :List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ :Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :int = feat_extract( UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) UpperCamelCase__ :Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) UpperCamelCase__ :Optional[int] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :Any = feat_extract( UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) UpperCamelCase__ :List[str] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase__ :Optional[Any] = np.random.rand(100 ).astype(np.floataa ) UpperCamelCase__ :Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCamelCase__ :List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) UpperCamelCase__ :List[str] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase__ :List[str] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase__ :Optional[int] = [np.asarray(UpperCamelCase_ ) for speech_input in speech_inputs] # Test feature size UpperCamelCase__ :Optional[int] = feature_extractor(audio_target=UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors='''np''' ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input UpperCamelCase__ :Union[str, Any] = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_values UpperCamelCase__ :Tuple = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3 ) ) # Test batched UpperCamelCase__ :int = feature_extractor(UpperCamelCase_ , return_tensors='''np''' ).input_values UpperCamelCase__ :Union[str, Any] = feature_extractor(UpperCamelCase_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_ ): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCamelCase__ :List[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCamelCase__ :Tuple = np.asarray(UpperCamelCase_ ) UpperCamelCase__ :int = feature_extractor(UpperCamelCase_ , return_tensors='''np''' ).input_values UpperCamelCase__ :List[str] = feature_extractor(UpperCamelCase_ , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_ ): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3 ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.feat_extract_tester.prepare_inputs_for_target() UpperCamelCase__ :str = self.feature_extraction_class(**self.feat_extract_dict ) UpperCamelCase__ :str = feat_extract.model_input_names[0] UpperCamelCase__ :int = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(UpperCamelCase_ ) == len(UpperCamelCase_ ) for x, y in zip(UpperCamelCase_ , processed_features[input_name] ) ) ) UpperCamelCase__ :List[str] = self.feat_extract_tester.prepare_inputs_for_target(equal_length=UpperCamelCase_ ) UpperCamelCase__ :Dict = BatchFeature({input_name: speech_inputs} , tensor_type='''np''' ) UpperCamelCase__ :List[str] = processed_features[input_name] if len(batch_features_input.shape ) < 3: UpperCamelCase__ :Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = self.feat_extract_tester.prepare_inputs_for_target(equal_length=UpperCamelCase_ ) UpperCamelCase__ :str = self.feature_extraction_class(**self.feat_extract_dict ) UpperCamelCase__ :Optional[int] = feat_extract.model_input_names[0] UpperCamelCase__ :str = BatchFeature({input_name: speech_inputs} , tensor_type='''pt''' ) UpperCamelCase__ :str = processed_features[input_name] if len(batch_features_input.shape ) < 3: UpperCamelCase__ :Tuple = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.feature_extraction_class(**self.feat_extract_dict ) UpperCamelCase__ :str = self.feat_extract_tester.prepare_inputs_for_target() UpperCamelCase__ :Any = feat_extract.model_input_names[0] UpperCamelCase__ :Union[str, Any] = BatchFeature({input_name: speech_inputs} ) UpperCamelCase__ :List[str] = feat_extract.num_mel_bins # hack! UpperCamelCase__ :int = feat_extract.pad(UpperCamelCase_ , padding='''longest''' , return_tensors='''np''' )[input_name] UpperCamelCase__ :List[str] = feat_extract.pad(UpperCamelCase_ , padding='''longest''' , return_tensors='''pt''' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.feat_extract_dict UpperCamelCase__ :int = True UpperCamelCase__ :Optional[int] = self.feature_extraction_class(**UpperCamelCase_ ) UpperCamelCase__ :int = self.feat_extract_tester.prepare_inputs_for_target() UpperCamelCase__ :List[str] = [len(UpperCamelCase_ ) for x in speech_inputs] UpperCamelCase__ :Dict = feat_extract.model_input_names[0] UpperCamelCase__ :Tuple = BatchFeature({input_name: speech_inputs} ) UpperCamelCase__ :Tuple = feat_extract.num_mel_bins # hack! UpperCamelCase__ :Union[str, Any] = feat_extract.pad(UpperCamelCase_ , padding='''longest''' , return_tensors='''np''' ) self.assertIn('''attention_mask''' , UpperCamelCase_ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.feat_extract_dict UpperCamelCase__ :int = True UpperCamelCase__ :Dict = self.feature_extraction_class(**UpperCamelCase_ ) UpperCamelCase__ :Union[str, Any] = self.feat_extract_tester.prepare_inputs_for_target() UpperCamelCase__ :List[Any] = [len(UpperCamelCase_ ) for x in speech_inputs] UpperCamelCase__ :List[str] = feat_extract.model_input_names[0] UpperCamelCase__ :Optional[Any] = BatchFeature({input_name: speech_inputs} ) UpperCamelCase__ :Dict = min(UpperCamelCase_ ) UpperCamelCase__ :Optional[int] = feat_extract.num_mel_bins # hack! UpperCamelCase__ :Optional[int] = feat_extract.pad( UpperCamelCase_ , padding='''max_length''' , max_length=UpperCamelCase_ , truncation=UpperCamelCase_ , return_tensors='''np''' ) self.assertIn('''attention_mask''' , UpperCamelCase_ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' from datasets import load_dataset UpperCamelCase__ :Union[str, Any] = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech UpperCamelCase__ :List[Any] = ds.sort('''id''' ).select(range(UpperCamelCase_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = torch.tensor( [2.38_04e-03, 2.07_52e-03, 1.98_36e-03, 2.10_57e-03, 1.61_74e-03, 3.05_18e-04, 9.15_53e-05, 3.35_69e-04, 9.76_56e-04, 1.83_11e-03, 2.01_42e-03, 2.10_57e-03, 1.73_95e-03, 4.57_76e-04, -3.96_73e-04, 4.57_76e-04, 1.00_71e-03, 9.15_53e-05, 4.88_28e-04, 1.15_97e-03, 7.32_42e-04, 9.46_04e-04, 1.80_05e-03, 1.83_11e-03, 8.85_01e-04, 4.27_25e-04, 4.88_28e-04, 7.32_42e-04, 1.09_86e-03, 2.10_57e-03] ) # fmt: on UpperCamelCase__ :List[str] = self._load_datasamples(1 ) UpperCamelCase__ :Any = SpeechTaFeatureExtractor() UpperCamelCase__ :Union[str, Any] = feature_extractor(UpperCamelCase_ , return_tensors='''pt''' ).input_values self.assertEquals(input_values.shape , (1, 93680) ) self.assertTrue(torch.allclose(input_values[0, :30] , UpperCamelCase_ , atol=1e-6 ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = torch.tensor( [-2.6870, -3.0104, -3.1356, -3.5352, -3.0044, -3.0353, -3.4719, -3.6777, -3.1520, -2.9435, -2.6553, -2.8795, -2.9944, -2.5921, -3.0279, -3.0386, -3.0864, -3.1291, -3.2353, -2.7444, -2.6831, -2.7287, -3.1761, -3.1571, -3.2726, -3.0582, -3.1007, -3.4533, -3.4695, -3.0998] ) # fmt: on UpperCamelCase__ :Optional[int] = self._load_datasamples(1 ) UpperCamelCase__ :str = SpeechTaFeatureExtractor() UpperCamelCase__ :Any = feature_extractor(audio_target=UpperCamelCase_ , return_tensors='''pt''' ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , UpperCamelCase_ , atol=1e-4 ) )
97
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Union[str, Any] = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE : Union[str, Any] = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE : Dict = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE : str = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE : Dict = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } SCREAMING_SNAKE_CASE : Optional[Any] = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } SCREAMING_SNAKE_CASE : Dict = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } SCREAMING_SNAKE_CASE : List[Any] = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } SCREAMING_SNAKE_CASE : Dict = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } SCREAMING_SNAKE_CASE : Dict = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class _lowerCamelCase( _a ): lowercase_ : Any = VOCAB_FILES_NAMES lowercase_ : Optional[int] = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowercase_ : str = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : str = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION class _lowerCamelCase( _a ): lowercase_ : Optional[int] = VOCAB_FILES_NAMES lowercase_ : Any = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowercase_ : str = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : Union[str, Any] = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : Optional[int] = collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) SCREAMING_SNAKE_CASE : Any = collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) SCREAMING_SNAKE_CASE : str = r"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n ```\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n ```\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Returns:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(_a ) class _lowerCamelCase: def __call__( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = False, lowerCamelCase = False, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, **lowerCamelCase, ) -> BatchEncoding: """simple docstring""" if titles is None and texts is None: return super().__call__( lowerCamelCase, padding=lowerCamelCase, truncation=lowerCamelCase, max_length=lowerCamelCase, return_tensors=lowerCamelCase, return_attention_mask=lowerCamelCase, **lowerCamelCase, ) elif titles is None or texts is None: _lowercase : Dict = titles if texts is None else texts return super().__call__( lowerCamelCase, lowerCamelCase, padding=lowerCamelCase, truncation=lowerCamelCase, max_length=lowerCamelCase, return_tensors=lowerCamelCase, return_attention_mask=lowerCamelCase, **lowerCamelCase, ) _lowercase : Union[str, Any] = titles if not isinstance(lowerCamelCase, lowerCamelCase) else [titles] _lowercase : Tuple = texts if not isinstance(lowerCamelCase, lowerCamelCase) else [texts] _lowercase : Optional[Any] = len(lowerCamelCase) _lowercase : Any = questions if not isinstance(lowerCamelCase, lowerCamelCase) else [questions] * n_passages if len(lowerCamelCase) != len(lowerCamelCase): raise ValueError( F'''There should be as many titles than texts but got {len(lowerCamelCase)} titles and {len(lowerCamelCase)} texts.''') _lowercase : Any = super().__call__(lowerCamelCase, lowerCamelCase, padding=lowerCamelCase, truncation=lowerCamelCase)['input_ids'] _lowercase : Tuple = super().__call__(lowerCamelCase, add_special_tokens=lowerCamelCase, padding=lowerCamelCase, truncation=lowerCamelCase)['input_ids'] _lowercase : int = { 'input_ids': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowerCamelCase, lowerCamelCase) ] } if return_attention_mask is not False: _lowercase : Optional[Any] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id) for input_id in input_ids]) _lowercase : Union[str, Any] = attention_mask return self.pad(lowerCamelCase, padding=lowerCamelCase, max_length=lowerCamelCase, return_tensors=lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = 16, lowerCamelCase = 64, lowerCamelCase = 4, ) -> List[DPRSpanPrediction]: """simple docstring""" _lowercase : Union[str, Any] = reader_input['input_ids'] _lowercase , _lowercase , _lowercase : Tuple = reader_output[:3] _lowercase : Tuple = len(lowerCamelCase) _lowercase : str = sorted(range(lowerCamelCase), reverse=lowerCamelCase, key=relevance_logits.__getitem__) _lowercase : List[DPRReaderOutput] = [] for doc_id in sorted_docs: _lowercase : str = list(input_ids[doc_id]) # assuming question & title information is at the beginning of the sequence _lowercase : Any = sequence_ids.index(self.sep_token_id, 2) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: _lowercase : List[Any] = sequence_ids.index(self.pad_token_id) else: _lowercase : List[str] = len(lowerCamelCase) _lowercase : Tuple = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len], end_logits=end_logits[doc_id][passage_offset:sequence_len], max_answer_length=lowerCamelCase, top_spans=lowerCamelCase, ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index], relevance_score=relevance_logits[doc_id], doc_id=lowerCamelCase, start_index=lowerCamelCase, end_index=lowerCamelCase, text=self.decode(sequence_ids[start_index : end_index + 1]), )) if len(lowerCamelCase) >= num_spans: break return nbest_spans_predictions[:num_spans] def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[DPRSpanPrediction]: """simple docstring""" _lowercase : str = [] for start_index, start_score in enumerate(lowerCamelCase): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length]): scores.append(((start_index, start_index + answer_length), start_score + end_score)) _lowercase : Dict = sorted(lowerCamelCase, key=lambda lowerCamelCase: x[1], reverse=lowerCamelCase) _lowercase : List[str] = [] for (start_index, end_index), score in scores: if start_index > end_index: raise ValueError(F'''Wrong span indices: [{start_index}:{end_index}]''') _lowercase : Dict = end_index - start_index + 1 if length > max_answer_length: raise ValueError(F'''Span is too long: {length} > {max_answer_length}''') if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals): continue chosen_span_intervals.append((start_index, end_index)) if len(lowerCamelCase) == top_spans: break return chosen_span_intervals @add_end_docstrings(_a ) class _lowerCamelCase( _a, _a ): lowercase_ : Union[str, Any] = VOCAB_FILES_NAMES lowercase_ : Any = READER_PRETRAINED_VOCAB_FILES_MAP lowercase_ : Dict = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : Optional[int] = READER_PRETRAINED_INIT_CONFIGURATION lowercase_ : str = ["""input_ids""", """attention_mask"""]
21
0
from typing import List from .keymap import KEYMAP, get_character def lowerCamelCase__ ( _lowercase ): '''simple docstring''' def decorator(_lowercase ): UpperCAmelCase_ : int = getattr(_lowercase , '''handle_key''' , [] ) handle += [key] setattr(_lowercase , '''handle_key''' , _lowercase ) return func return decorator def lowerCamelCase__ ( *_lowercase ): '''simple docstring''' def decorator(_lowercase ): UpperCAmelCase_ : Tuple = getattr(_lowercase , '''handle_key''' , [] ) handle += keys setattr(_lowercase , '''handle_key''' , _lowercase ) return func return decorator class __a( _a ): """simple docstring""" def __new__( cls ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Dict: UpperCAmelCase_ : List[str] = super().__new__(cls ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) if not hasattr(_SCREAMING_SNAKE_CASE ,'''key_handler''' ): setattr(_SCREAMING_SNAKE_CASE ,'''key_handler''' ,{} ) setattr(_SCREAMING_SNAKE_CASE ,'''handle_input''' ,KeyHandler.handle_input ) for value in attrs.values(): UpperCAmelCase_ : str = getattr(_SCREAMING_SNAKE_CASE ,'''handle_key''' ,[] ) for key in handled_keys: UpperCAmelCase_ : Optional[Any] = value return new_cls @staticmethod def a__ ( cls ) -> Tuple: UpperCAmelCase_ : Union[str, Any] = get_character() if char != KEYMAP["undefined"]: UpperCAmelCase_ : Any = ord(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = cls.key_handler.get(_SCREAMING_SNAKE_CASE ) if handler: UpperCAmelCase_ : str = char return handler(cls ) else: return None def lowerCamelCase__ ( cls ): '''simple docstring''' return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
235
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __a = { 'configuration_bloom': ['BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BloomConfig', 'BloomOnnxConfig'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['BloomTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST', 'BloomForCausalLM', 'BloomModel', 'BloomPreTrainedModel', 'BloomForSequenceClassification', 'BloomForTokenClassification', 'BloomForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
235
1
'''simple docstring''' import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class _lowercase ( _lowercase ): a = (DDPMParallelScheduler,) def lowerCamelCase_ ( self: Union[str, Any] , **UpperCamelCase__: str ): lowerCamelCase__ : str = { """num_train_timesteps""": 1_000, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", """variance_type""": """fixed_small""", """clip_sample""": True, } config.update(**UpperCamelCase__ ) return config def lowerCamelCase_ ( self: Tuple ): for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=UpperCamelCase__ , beta_end=UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.check_over_configs(thresholding=UpperCamelCase__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=UpperCamelCase__ , prediction_type=UpperCamelCase__ , sample_max_value=UpperCamelCase__ , ) def lowerCamelCase_ ( self: str ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): for t in [0, 500, 999]: self.check_over_forward(time_step=UpperCamelCase__ ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : int = self.scheduler_classes[0] lowerCamelCase__ : Union[str, Any] = self.get_scheduler_config() lowerCamelCase__ : Union[str, Any] = scheduler_class(**UpperCamelCase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.00_979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1e-5 def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = self.scheduler_classes[0] lowerCamelCase__ : List[Any] = self.get_scheduler_config() lowerCamelCase__ : List[str] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : str = len(UpperCamelCase__ ) lowerCamelCase__ : str = self.dummy_model() lowerCamelCase__ : int = self.dummy_sample_deter lowerCamelCase__ : Optional[int] = self.dummy_sample_deter + 0.1 lowerCamelCase__ : Optional[int] = self.dummy_sample_deter - 0.1 lowerCamelCase__ : Union[str, Any] = samplea.shape[0] lowerCamelCase__ : Union[str, Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCamelCase__ : str = torch.arange(UpperCamelCase__ )[0:3, None].repeat(1 , UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCamelCase__ : Dict = scheduler.batch_step_no_noise(UpperCamelCase__ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCamelCase__ : Union[str, Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 1_153.1_833 ) < 1e-2 assert abs(result_mean.item() - 0.5_005 ) < 1e-3 def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[Any] = self.scheduler_classes[0] lowerCamelCase__ : Dict = self.get_scheduler_config() lowerCamelCase__ : List[Any] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = self.dummy_model() lowerCamelCase__ : int = self.dummy_sample_deter lowerCamelCase__ : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(UpperCamelCase__ ) ): # 1. predict noise residual lowerCamelCase__ : Dict = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 lowerCamelCase__ : Any = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample lowerCamelCase__ : List[str] = pred_prev_sample lowerCamelCase__ : List[Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : Optional[int] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 258.9_606 ) < 1e-2 assert abs(result_mean.item() - 0.3_372 ) < 1e-3 def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCamelCase__ : Any = self.get_scheduler_config(prediction_type="""v_prediction""" ) lowerCamelCase__ : Any = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : int = len(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = self.dummy_model() lowerCamelCase__ : List[Any] = self.dummy_sample_deter lowerCamelCase__ : List[str] = torch.manual_seed(0 ) for t in reversed(range(UpperCamelCase__ ) ): # 1. predict noise residual lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 lowerCamelCase__ : List[str] = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample lowerCamelCase__ : List[Any] = pred_prev_sample lowerCamelCase__ : Union[str, Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 202.0_296 ) < 1e-2 assert abs(result_mean.item() - 0.2_631 ) < 1e-3 def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Dict = self.scheduler_classes[0] lowerCamelCase__ : List[str] = self.get_scheduler_config() lowerCamelCase__ : Optional[int] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=UpperCamelCase__ ) lowerCamelCase__ : Any = scheduler.timesteps for i, timestep in enumerate(UpperCamelCase__ ): if i == len(UpperCamelCase__ ) - 1: lowerCamelCase__ : List[str] = -1 else: lowerCamelCase__ : int = timesteps[i + 1] lowerCamelCase__ : List[Any] = scheduler.previous_timestep(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = prev_t.item() self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Optional[int] = self.scheduler_classes[0] lowerCamelCase__ : Union[str, Any] = self.get_scheduler_config() lowerCamelCase__ : Any = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = [100, 87, 50, 51, 0] with self.assertRaises(UpperCamelCase__ , msg="""`custom_timesteps` must be in descending order.""" ): scheduler.set_timesteps(timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Tuple = self.scheduler_classes[0] lowerCamelCase__ : Dict = self.get_scheduler_config() lowerCamelCase__ : str = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : Dict = [100, 87, 50, 1, 0] lowerCamelCase__ : List[str] = len(UpperCamelCase__ ) with self.assertRaises(UpperCamelCase__ , msg="""Can only pass one of `num_inference_steps` or `custom_timesteps`.""" ): scheduler.set_timesteps(num_inference_steps=UpperCamelCase__ , timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCamelCase__ : Tuple = self.get_scheduler_config() lowerCamelCase__ : List[Any] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [scheduler.config.num_train_timesteps] with self.assertRaises( UpperCamelCase__ , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=UpperCamelCase__ )
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'post_extract_proj': 'feature_projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.upsample.0': 'encoder.upsample.projection', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'layer_norm', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: List[str] , lowerCAmelCase: str , lowerCAmelCase: Optional[int] , lowerCAmelCase: Union[str, Any] , lowerCAmelCase: Union[str, Any] ) -> Optional[int]: for attribute in key.split("." ): _UpperCAmelCase : Tuple = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) if weight_type is not None: _UpperCAmelCase : Optional[int] = getattr(lowerCAmelCase__ , lowerCAmelCase__ ).shape else: _UpperCAmelCase : str = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _UpperCAmelCase : Tuple = value elif weight_type == "weight_g": _UpperCAmelCase : str = value elif weight_type == "weight_v": _UpperCAmelCase : List[str] = value elif weight_type == "bias": _UpperCAmelCase : Tuple = value else: _UpperCAmelCase : Union[str, Any] = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: str , lowerCAmelCase: Any , lowerCAmelCase: Optional[Any] ) -> Any: _UpperCAmelCase : Any = [] _UpperCAmelCase : Dict = fairseq_model.state_dict() _UpperCAmelCase : Union[str, Any] = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _UpperCAmelCase : Optional[Any] = False if "conv_layers" in name: load_conv_layer( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , hf_model.config.feat_extract_norm == "group" , ) _UpperCAmelCase : Union[str, Any] = True else: for key, mapped_key in MAPPING.items(): _UpperCAmelCase : List[Any] = """sew.""" + mapped_key if (is_finetuned and mapped_key != """lm_head""") else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: _UpperCAmelCase : int = True if "*" in mapped_key: _UpperCAmelCase : Dict = name.split(lowerCAmelCase__ )[0].split("." )[-2] _UpperCAmelCase : List[str] = mapped_key.replace("*" , lowerCAmelCase__ ) if "weight_g" in name: _UpperCAmelCase : List[str] = """weight_g""" elif "weight_v" in name: _UpperCAmelCase : List[str] = """weight_v""" elif "weight" in name: _UpperCAmelCase : int = """weight""" elif "bias" in name: _UpperCAmelCase : Optional[int] = """bias""" else: _UpperCAmelCase : List[str] = None set_recursively(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) continue if not is_used: unused_weights.append(lowerCAmelCase__ ) logger.warning(F'Unused weights: {unused_weights}' ) def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: int , lowerCAmelCase: Tuple , lowerCAmelCase: str , lowerCAmelCase: Tuple , lowerCAmelCase: Any ) -> Any: _UpperCAmelCase : Tuple = full_name.split("conv_layers." )[-1] _UpperCAmelCase : Tuple = name.split("." ) _UpperCAmelCase : List[Any] = int(items[0] ) _UpperCAmelCase : Optional[int] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _UpperCAmelCase : Dict = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _UpperCAmelCase : Union[str, Any] = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _UpperCAmelCase : Union[str, Any] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _UpperCAmelCase : List[str] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(lowerCAmelCase__ ) def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: Tuple , lowerCAmelCase: Dict ) -> int: _UpperCAmelCase : Dict = SEWConfig() if is_finetuned: _UpperCAmelCase : Optional[int] = model.wav_encoder.wav_model.cfg else: _UpperCAmelCase : int = model.cfg _UpperCAmelCase : Union[str, Any] = fs_config.conv_bias _UpperCAmelCase : Optional[Any] = eval(fs_config.conv_feature_layers ) _UpperCAmelCase : Any = [x[0] for x in conv_layers] _UpperCAmelCase : str = [x[1] for x in conv_layers] _UpperCAmelCase : Union[str, Any] = [x[2] for x in conv_layers] _UpperCAmelCase : Optional[int] = """gelu""" _UpperCAmelCase : Optional[Any] = """layer""" if fs_config.extractor_mode == """layer_norm""" else """group""" _UpperCAmelCase : List[Any] = 0.0 _UpperCAmelCase : int = fs_config.activation_fn.name _UpperCAmelCase : Optional[Any] = fs_config.encoder_embed_dim _UpperCAmelCase : List[Any] = 0.02 _UpperCAmelCase : Optional[int] = fs_config.encoder_ffn_embed_dim _UpperCAmelCase : List[str] = 1E-5 _UpperCAmelCase : str = fs_config.encoder_layerdrop _UpperCAmelCase : str = fs_config.encoder_attention_heads _UpperCAmelCase : List[str] = fs_config.conv_pos_groups _UpperCAmelCase : int = fs_config.conv_pos _UpperCAmelCase : int = len(lowerCAmelCase__ ) _UpperCAmelCase : Any = fs_config.encoder_layers _UpperCAmelCase : List[Any] = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _UpperCAmelCase : int = model.cfg _UpperCAmelCase : Any = fs_config.final_dropout _UpperCAmelCase : Union[str, Any] = fs_config.layerdrop _UpperCAmelCase : Dict = fs_config.activation_dropout _UpperCAmelCase : List[Any] = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _UpperCAmelCase : Optional[Any] = fs_config.attention_dropout _UpperCAmelCase : Dict = fs_config.dropout_input _UpperCAmelCase : Optional[int] = fs_config.dropout _UpperCAmelCase : Dict = fs_config.mask_channel_length _UpperCAmelCase : Optional[int] = fs_config.mask_channel_prob _UpperCAmelCase : Union[str, Any] = fs_config.mask_length _UpperCAmelCase : List[Any] = fs_config.mask_prob _UpperCAmelCase : Dict = """Wav2Vec2FeatureExtractor""" _UpperCAmelCase : Dict = """Wav2Vec2CTCTokenizer""" return config @torch.no_grad() def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: Optional[int] , lowerCAmelCase: Dict , lowerCAmelCase: Optional[int]=None , lowerCAmelCase: List[str]=None , lowerCAmelCase: List[Any]=True ) -> Union[str, Any]: if is_finetuned: _UpperCAmelCase : List[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) else: _UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _UpperCAmelCase : Optional[Any] = SEWConfig.from_pretrained(lowerCAmelCase__ ) else: _UpperCAmelCase : int = convert_config(model[0] , lowerCAmelCase__ ) _UpperCAmelCase : str = model[0].eval() _UpperCAmelCase : Tuple = True if config.feat_extract_norm == """layer""" else False _UpperCAmelCase : Union[str, Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , ) if is_finetuned: if dict_path: _UpperCAmelCase : Union[str, Any] = Dictionary.load(lowerCAmelCase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _UpperCAmelCase : Dict = target_dict.pad_index _UpperCAmelCase : Tuple = target_dict.bos_index _UpperCAmelCase : Optional[int] = target_dict.pad_index _UpperCAmelCase : str = target_dict.bos_index _UpperCAmelCase : List[Any] = target_dict.eos_index _UpperCAmelCase : List[str] = len(target_dict.symbols ) _UpperCAmelCase : Dict = os.path.join(lowerCAmelCase__ , "vocab.json" ) if not os.path.isdir(lowerCAmelCase__ ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(lowerCAmelCase__ ) ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as vocab_handle: json.dump(target_dict.indices , lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = WavaVecaCTCTokenizer( lowerCAmelCase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=lowerCAmelCase__ , ) _UpperCAmelCase : Optional[Any] = WavaVecaProcessor(feature_extractor=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) processor.save_pretrained(lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = SEWForCTC(lowerCAmelCase__ ) else: _UpperCAmelCase : List[Any] = SEWModel(lowerCAmelCase__ ) feature_extractor.save_pretrained(lowerCAmelCase__ ) recursively_load_weights(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) hf_model.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--is_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) SCREAMING_SNAKE_CASE_ = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
366
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) # General docstring SCREAMING_SNAKE_CASE_ = 'ResNetConfig' # Base docstring SCREAMING_SNAKE_CASE_ = 'microsoft/resnet-50' SCREAMING_SNAKE_CASE_ = [1, 2048, 7, 7] # Image classification docstring SCREAMING_SNAKE_CASE_ = 'microsoft/resnet-50' SCREAMING_SNAKE_CASE_ = 'tiger cat' SCREAMING_SNAKE_CASE_ = [ 'microsoft/resnet-50', # See all resnet models at https://huggingface.co/models?filter=resnet ] class a ( nn.Module ): def __init__( self , A_ , A_ , A_ = 3 , A_ = 1 , A_ = "relu" ): '''simple docstring''' super().__init__() _UpperCAmelCase : Union[str, Any] = nn.Convad( A_ , A_ , kernel_size=A_ , stride=A_ , padding=kernel_size // 2 , bias=A_ ) _UpperCAmelCase : List[Any] = nn.BatchNormad(A_ ) _UpperCAmelCase : Union[str, Any] = ACTaFN[activation] if activation is not None else nn.Identity() def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : List[Any] = self.convolution(A_ ) _UpperCAmelCase : Optional[int] = self.normalization(A_ ) _UpperCAmelCase : Optional[Any] = self.activation(A_ ) return hidden_state class a ( nn.Module ): def __init__( self , A_ ): '''simple docstring''' super().__init__() _UpperCAmelCase : Any = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) _UpperCAmelCase : List[str] = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) _UpperCAmelCase : List[Any] = config.num_channels def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : int = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) _UpperCAmelCase : int = self.embedder(A_ ) _UpperCAmelCase : int = self.pooler(A_ ) return embedding class a ( nn.Module ): def __init__( self , A_ , A_ , A_ = 2 ): '''simple docstring''' super().__init__() _UpperCAmelCase : Union[str, Any] = nn.Convad(A_ , A_ , kernel_size=1 , stride=A_ , bias=A_ ) _UpperCAmelCase : Optional[int] = nn.BatchNormad(A_ ) def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : str = self.convolution(A_ ) _UpperCAmelCase : List[str] = self.normalization(A_ ) return hidden_state class a ( nn.Module ): def __init__( self , A_ , A_ , A_ = 1 , A_ = "relu" ): '''simple docstring''' super().__init__() _UpperCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Dict = ( ResNetShortCut(A_ , A_ , stride=A_ ) if should_apply_shortcut else nn.Identity() ) _UpperCAmelCase : int = nn.Sequential( ResNetConvLayer(A_ , A_ , stride=A_ ) , ResNetConvLayer(A_ , A_ , activation=A_ ) , ) _UpperCAmelCase : Dict = ACTaFN[activation] def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : Optional[Any] = hidden_state _UpperCAmelCase : Any = self.layer(A_ ) _UpperCAmelCase : Optional[int] = self.shortcut(A_ ) hidden_state += residual _UpperCAmelCase : Optional[int] = self.activation(A_ ) return hidden_state class a ( nn.Module ): def __init__( self , A_ , A_ , A_ = 1 , A_ = "relu" , A_ = 4 ): '''simple docstring''' super().__init__() _UpperCAmelCase : Optional[Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = out_channels // reduction _UpperCAmelCase : List[str] = ( ResNetShortCut(A_ , A_ , stride=A_ ) if should_apply_shortcut else nn.Identity() ) _UpperCAmelCase : Dict = nn.Sequential( ResNetConvLayer(A_ , A_ , kernel_size=1 ) , ResNetConvLayer(A_ , A_ , stride=A_ ) , ResNetConvLayer(A_ , A_ , kernel_size=1 , activation=A_ ) , ) _UpperCAmelCase : List[str] = ACTaFN[activation] def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : List[Any] = hidden_state _UpperCAmelCase : List[str] = self.layer(A_ ) _UpperCAmelCase : List[str] = self.shortcut(A_ ) hidden_state += residual _UpperCAmelCase : Dict = self.activation(A_ ) return hidden_state class a ( nn.Module ): def __init__( self , A_ , A_ , A_ , A_ = 2 , A_ = 2 , ): '''simple docstring''' super().__init__() _UpperCAmelCase : Any = ResNetBottleNeckLayer if config.layer_type == "bottleneck" else ResNetBasicLayer _UpperCAmelCase : Optional[Any] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(A_ , A_ , stride=A_ , activation=config.hidden_act ) , *[layer(A_ , A_ , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : List[Any] = input for layer in self.layers: _UpperCAmelCase : Optional[Any] = layer(A_ ) return hidden_state class a ( nn.Module ): def __init__( self , A_ ): '''simple docstring''' super().__init__() _UpperCAmelCase : Any = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( A_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) _UpperCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(A_ , config.depths[1:] ): self.stages.append(ResNetStage(A_ , A_ , A_ , depth=A_ ) ) def _UpperCAmelCase ( self , A_ , A_ = False , A_ = True ): '''simple docstring''' _UpperCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Dict = hidden_states + (hidden_state,) _UpperCAmelCase : str = stage_module(A_ ) if output_hidden_states: _UpperCAmelCase : int = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=A_ , hidden_states=A_ , ) class a ( UpperCAmelCase ): _lowercase = ResNetConfig _lowercase = "resnet" _lowercase = "pixel_values" _lowercase = True def _UpperCAmelCase ( self , A_ ): '''simple docstring''' if isinstance(A_ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu" ) elif isinstance(A_ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _UpperCAmelCase ( self , A_ , A_=False ): '''simple docstring''' if isinstance(A_ , A_ ): _UpperCAmelCase : Optional[Any] = value SCREAMING_SNAKE_CASE_ = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`ResNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' SCREAMING_SNAKE_CASE_ = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare ResNet model outputting raw features without any specific head on top." , UpperCAmelCase , ) class a ( UpperCAmelCase ): def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ ) _UpperCAmelCase : List[str] = config _UpperCAmelCase : Any = ResNetEmbeddings(A_ ) _UpperCAmelCase : str = ResNetEncoder(A_ ) _UpperCAmelCase : Any = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A_ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self , A_ , A_ = None , A_ = None ): '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : List[Any] = self.embedder(A_ ) _UpperCAmelCase : str = self.encoder( A_ , output_hidden_states=A_ , return_dict=A_ ) _UpperCAmelCase : List[Any] = encoder_outputs[0] _UpperCAmelCase : int = self.pooler(A_ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A_ , pooler_output=A_ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( "\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase , ) class a ( UpperCAmelCase ): def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : str = ResNetModel(A_ ) # classification head _UpperCAmelCase : int = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self , A_ = None , A_ = None , A_ = None , A_ = None , ): '''simple docstring''' _UpperCAmelCase : Optional[int] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Tuple = self.resnet(A_ , output_hidden_states=A_ , return_dict=A_ ) _UpperCAmelCase : Optional[int] = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : int = self.classifier(A_ ) _UpperCAmelCase : Dict = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _UpperCAmelCase : Optional[Any] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _UpperCAmelCase : Optional[Any] = "single_label_classification" else: _UpperCAmelCase : Any = "multi_label_classification" if self.config.problem_type == "regression": _UpperCAmelCase : str = MSELoss() if self.num_labels == 1: _UpperCAmelCase : Any = loss_fct(logits.squeeze() , labels.squeeze() ) else: _UpperCAmelCase : Optional[int] = loss_fct(A_ , A_ ) elif self.config.problem_type == "single_label_classification": _UpperCAmelCase : Any = CrossEntropyLoss() _UpperCAmelCase : Any = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _UpperCAmelCase : Any = BCEWithLogitsLoss() _UpperCAmelCase : Tuple = loss_fct(A_ , A_ ) if not return_dict: _UpperCAmelCase : Any = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=A_ , logits=A_ , hidden_states=outputs.hidden_states ) @add_start_docstrings( "\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n " , UpperCAmelCase , ) class a ( UpperCAmelCase , UpperCAmelCase ): def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ ) super()._init_backbone(A_ ) _UpperCAmelCase : Optional[int] = [config.embedding_size] + config.hidden_sizes _UpperCAmelCase : str = ResNetEmbeddings(A_ ) _UpperCAmelCase : List[Any] = ResNetEncoder(A_ ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_ ) @replace_return_docstrings(output_type=A_ , config_class=_CONFIG_FOR_DOC ) def _UpperCAmelCase ( self , A_ , A_ = None , A_ = None ): '''simple docstring''' _UpperCAmelCase : Tuple = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Tuple = self.embedder(A_ ) _UpperCAmelCase : Optional[int] = self.encoder(A_ , output_hidden_states=A_ , return_dict=A_ ) _UpperCAmelCase : Optional[int] = outputs.hidden_states _UpperCAmelCase : Any = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: _UpperCAmelCase : Union[str, Any] = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=A_ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=A_ , )
189
0
import unittest import numpy as np def _SCREAMING_SNAKE_CASE ( a , a , a , a = None , ) -> np.ndarray: __A : List[str] = np.shape(a ) __A : Union[str, Any] = np.shape(a ) __A : str = np.shape(a ) if shape_a[0] != shape_b[0]: __A : Dict = ( 'Expected the same number of rows for A and B. ' F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(a ) if shape_b[1] != shape_c[1]: __A : Any = ( 'Expected the same number of columns for B and C. ' F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(a ) __A : str = pseudo_inv if a_inv is None: try: __A : Any = np.linalg.inv(a ) except np.linalg.LinAlgError: raise ValueError( 'Input matrix A is not invertible. Cannot compute Schur complement.' ) return mat_c - mat_b.T @ a_inv @ mat_b class _A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): __A : int = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __A : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) __A : str = np.array([[2, 1], [6, 3]] ) __A : Any = schur_complement(_A , _A , _A ) __A : int = np.block([[a, b], [b.T, c]] ) __A : List[Any] = np.linalg.det(_A ) __A : Optional[Any] = np.linalg.det(_A ) __A : List[str] = np.linalg.det(_A ) self.assertAlmostEqual(_A , det_a * det_s ) def UpperCAmelCase_ ( self ): __A : Optional[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __A : Any = np.array([[0, 3], [3, 0], [2, 3]] ) __A : Tuple = np.array([[2, 1], [6, 3]] ) with self.assertRaises(_A ): schur_complement(_A , _A , _A ) def UpperCAmelCase_ ( self ): __A : List[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __A : int = np.array([[0, 3], [3, 0], [2, 3]] ) __A : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(_A ): schur_complement(_A , _A , _A ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
280
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
1
_lowerCamelCase : Optional[Any] = 2_5_6 # Modulus to hash a string _lowerCamelCase : Tuple = 1_0_0_0_0_0_3 def a__ ( UpperCAmelCase : str , UpperCAmelCase : str ) -> bool: UpperCAmelCase : Dict = len(UpperCAmelCase ) UpperCAmelCase : Any = len(UpperCAmelCase ) if p_len > t_len: return False UpperCAmelCase : Union[str, Any] = 0 UpperCAmelCase : Any = 0 UpperCAmelCase : List[str] = 1 # Calculating the hash of pattern and substring of text for i in range(UpperCAmelCase ): UpperCAmelCase : Any = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus UpperCAmelCase : List[str] = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue UpperCAmelCase : Optional[Any] = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash UpperCAmelCase : Any = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def a__ ( ) -> None: UpperCAmelCase : Optional[Any] = '''abc1abc12''' UpperCAmelCase : List[str] = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' UpperCAmelCase : Tuple = '''alskfjaldsk23adsfabcabc''' assert rabin_karp(UpperCAmelCase , UpperCAmelCase ) and not rabin_karp(UpperCAmelCase , UpperCAmelCase ) # Test 2) UpperCAmelCase : int = '''ABABX''' UpperCAmelCase : Dict = '''ABABZABABYABABX''' assert rabin_karp(UpperCAmelCase , UpperCAmelCase ) # Test 3) UpperCAmelCase : int = '''AAAB''' UpperCAmelCase : List[Any] = '''ABAAAAAB''' assert rabin_karp(UpperCAmelCase , UpperCAmelCase ) # Test 4) UpperCAmelCase : Optional[int] = '''abcdabcy''' UpperCAmelCase : List[str] = '''abcxabcdabxabcdabcdabcy''' assert rabin_karp(UpperCAmelCase , UpperCAmelCase ) # Test 5) UpperCAmelCase : int = '''Lü''' UpperCAmelCase : int = '''Lüsai''' assert rabin_karp(UpperCAmelCase , UpperCAmelCase ) UpperCAmelCase : str = '''Lue''' assert not rabin_karp(UpperCAmelCase , UpperCAmelCase ) print('''Success.''' ) if __name__ == "__main__": test_rabin_karp()
354
from math import log from scipy.constants import Boltzmann, physical_constants _lowerCamelCase : Tuple = 3_0_0 # TEMPERATURE (unit = K) def a__ ( UpperCAmelCase : float , UpperCAmelCase : float , UpperCAmelCase : float , ) -> float: if donor_conc <= 0: raise ValueError('''Donor concentration should be positive''' ) elif acceptor_conc <= 0: raise ValueError('''Acceptor concentration should be positive''' ) elif intrinsic_conc <= 0: raise ValueError('''Intrinsic concentration should be positive''' ) elif donor_conc <= intrinsic_conc: raise ValueError( '''Donor concentration should be greater than intrinsic concentration''' ) elif acceptor_conc <= intrinsic_conc: raise ValueError( '''Acceptor concentration should be greater than intrinsic concentration''' ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
99
0