code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : int = IFInpaintingPipeline __UpperCAmelCase : int = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} __UpperCAmelCase : Dict = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCAmelCase : Union[str, Any] = PipelineTesterMixin.required_optional_params - {"latents"} def __snake_case ( self : int ) -> str: return self._get_dummy_components() def __snake_case ( self : Any , lowerCamelCase : Dict , lowerCamelCase : Dict=0 ) -> Dict: if str(lowerCamelCase ).startswith("mps" ): __snake_case : Optional[Any] = torch.manual_seed(lowerCamelCase ) else: __snake_case : Any = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __snake_case : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __snake_case : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __snake_case : Union[str, Any] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __snake_case ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self : Union[str, Any] ) -> List[str]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __snake_case ( self : Optional[int] ) -> str: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self : Any ) -> Any: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self : List[Any] ) -> Tuple: self._test_save_load_local() def __snake_case ( self : List[Any] ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
81
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
0
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCamelCase = (720, 1_280) # Height, Width lowerCamelCase = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCamelCase = 1 / 100 lowerCamelCase = """""" lowerCamelCase = """""" lowerCamelCase = """""" lowerCamelCase = 250 def a__ ( ): UpperCAmelCase_ , UpperCAmelCase_ = get_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) for index in range(lowerCAmelCase__ ): UpperCAmelCase_ = random.sample(range(len(lowerCAmelCase__ ) ) , 4 ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = update_image_and_anno( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , filter_scale=lowerCAmelCase__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' UpperCAmelCase_ = random_chars(32 ) UpperCAmelCase_ = path.split(os.sep )[-1].rsplit("." , 1 )[0] UpperCAmelCase_ = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , lowerCAmelCase__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) UpperCAmelCase_ = [] for anno in new_annos: UpperCAmelCase_ = anno[3] - anno[1] UpperCAmelCase_ = anno[4] - anno[2] UpperCAmelCase_ = anno[1] + width / 2 UpperCAmelCase_ = anno[2] + height / 2 UpperCAmelCase_ = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(lowerCAmelCase__ ) with open(f"""{file_root}.txt""" , "w" ) as outfile: outfile.write("\n".join(line for line in annos_list ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = [] UpperCAmelCase_ = [] for label_file in glob.glob(os.path.join(lowerCAmelCase__ , "*.txt" ) ): UpperCAmelCase_ = label_file.split(os.sep )[-1].rsplit("." , 1 )[0] with open(lowerCAmelCase__ ) as in_file: UpperCAmelCase_ = in_file.readlines() UpperCAmelCase_ = os.path.join(lowerCAmelCase__ , f"""{label_name}.jpg""" ) UpperCAmelCase_ = [] for obj_list in obj_lists: UpperCAmelCase_ = obj_list.rstrip("\n" ).split(" " ) UpperCAmelCase_ = float(obj[1] ) - float(obj[3] ) / 2 UpperCAmelCase_ = float(obj[2] ) - float(obj[4] ) / 2 UpperCAmelCase_ = float(obj[1] ) + float(obj[3] ) / 2 UpperCAmelCase_ = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(lowerCAmelCase__ ) labels.append(lowerCAmelCase__ ) return img_paths, labels def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , ): UpperCAmelCase_ = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) UpperCAmelCase_ = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCAmelCase_ = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCAmelCase_ = int(scale_x * output_size[1] ) UpperCAmelCase_ = int(scale_y * output_size[0] ) UpperCAmelCase_ = [] UpperCAmelCase_ = [] for i, index in enumerate(lowerCAmelCase__ ): UpperCAmelCase_ = all_img_list[index] path_list.append(lowerCAmelCase__ ) UpperCAmelCase_ = all_annos[index] UpperCAmelCase_ = cva.imread(lowerCAmelCase__ ) if i == 0: # top-left UpperCAmelCase_ = cva.resize(lowerCAmelCase__ , (divid_point_x, divid_point_y) ) UpperCAmelCase_ = img for bbox in img_annos: UpperCAmelCase_ = bbox[1] * scale_x UpperCAmelCase_ = bbox[2] * scale_y UpperCAmelCase_ = bbox[3] * scale_x UpperCAmelCase_ = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right UpperCAmelCase_ = cva.resize(lowerCAmelCase__ , (output_size[1] - divid_point_x, divid_point_y) ) UpperCAmelCase_ = img for bbox in img_annos: UpperCAmelCase_ = scale_x + bbox[1] * (1 - scale_x) UpperCAmelCase_ = bbox[2] * scale_y UpperCAmelCase_ = scale_x + bbox[3] * (1 - scale_x) UpperCAmelCase_ = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left UpperCAmelCase_ = cva.resize(lowerCAmelCase__ , (divid_point_x, output_size[0] - divid_point_y) ) UpperCAmelCase_ = img for bbox in img_annos: UpperCAmelCase_ = bbox[1] * scale_x UpperCAmelCase_ = scale_y + bbox[2] * (1 - scale_y) UpperCAmelCase_ = bbox[3] * scale_x UpperCAmelCase_ = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right UpperCAmelCase_ = cva.resize( lowerCAmelCase__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) UpperCAmelCase_ = img for bbox in img_annos: UpperCAmelCase_ = scale_x + bbox[1] * (1 - scale_x) UpperCAmelCase_ = scale_y + bbox[2] * (1 - scale_y) UpperCAmelCase_ = scale_x + bbox[3] * (1 - scale_x) UpperCAmelCase_ = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: UpperCAmelCase_ = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def a__ ( lowerCAmelCase__ ): assert number_char > 1, "The number of character should greater than 1" UpperCAmelCase_ = ascii_lowercase + digits return "".join(random.choice(lowerCAmelCase__ ) for _ in range(lowerCAmelCase__ ) ) if __name__ == "__main__": main() print("""DONE ✅""")
82
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
0
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case_ ( A_ : Tuple, A_ : List[str], A_ : Optional[Any], A_ : Dict, A_ : Dict=True, A_ : int="pt" ): '''simple docstring''' _lowerCamelCase : str = {'''add_prefix_space''': True} if isinstance(A_, A_ ) and not line.startswith(''' ''' ) else {} _lowerCamelCase : Union[str, Any] = padding_side return tokenizer( [line], max_length=A_, padding='''max_length''' if pad_to_max_length else None, truncation=A_, return_tensors=A_, add_special_tokens=A_, **A_, ) def snake_case_ ( A_ : Any, A_ : Optional[int], A_ : List[Any]=None, ): '''simple docstring''' _lowerCamelCase : Optional[int] = input_ids.ne(A_ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __snake_case ( _lowercase): def __init__( self : Tuple , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Tuple="train" , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[int]=None , __lowerCAmelCase : Any=None , __lowerCAmelCase : Union[str, Any]="" , ): """simple docstring""" super().__init__() _lowerCamelCase : Optional[int] = Path(__lowerCAmelCase ).joinpath(type_path + '''.source''' ) _lowerCamelCase : List[str] = Path(__lowerCAmelCase ).joinpath(type_path + '''.target''' ) _lowerCamelCase : List[Any] = self.get_char_lens(self.src_file ) _lowerCamelCase : Optional[int] = max_source_length _lowerCamelCase : Optional[Any] = max_target_length assert min(self.src_lens ) > 0, f'''found empty line in {self.src_file}''' _lowerCamelCase : List[Any] = tokenizer _lowerCamelCase : List[Any] = prefix if n_obs is not None: _lowerCamelCase : List[str] = self.src_lens[:n_obs] _lowerCamelCase : int = src_lang _lowerCamelCase : Union[str, Any] = tgt_lang def __len__( self : int ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : Dict , __lowerCAmelCase : Optional[Any] ): """simple docstring""" _lowerCamelCase : str = index + 1 # linecache starts at 1 _lowerCamelCase : Union[str, Any] = self.prefix + linecache.getline(str(self.src_file ) , __lowerCAmelCase ).rstrip('''\n''' ) _lowerCamelCase : Optional[Any] = linecache.getline(str(self.tgt_file ) , __lowerCAmelCase ).rstrip('''\n''' ) assert source_line, f'''empty source line for index {index}''' assert tgt_line, f'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer , __lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right _lowerCamelCase : Optional[int] = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer ) _lowerCamelCase : Union[str, Any] = self.tokenizer.generator if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_source_length , '''right''' ) _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_target_length , '''right''' ) _lowerCamelCase : Optional[Any] = source_inputs['''input_ids'''].squeeze() _lowerCamelCase : Union[str, Any] = target_inputs['''input_ids'''].squeeze() _lowerCamelCase : Any = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : str ): """simple docstring""" return [len(__lowerCAmelCase ) for x in Path(__lowerCAmelCase ).open().readlines()] def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[Any] = torch.stack([x['''input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = torch.stack([x['''attention_mask'''] for x in batch] ) _lowerCamelCase : Union[str, Any] = torch.stack([x['''decoder_input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Tuple = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Union[str, Any] = trim_batch(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : List[str] = trim_batch(__lowerCAmelCase , __lowerCAmelCase , attention_mask=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch lowerCAmelCase__ = getLogger(__name__) def snake_case_ ( A_ : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(A_ ) ) def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : Dict = get_git_info() save_json(A_, os.path.join(A_, '''git_log.json''' ) ) def snake_case_ ( A_ : str, A_ : Union[str, Any], A_ : int=4, **A_ : Optional[int] ): '''simple docstring''' with open(A_, '''w''' ) as f: json.dump(A_, A_, indent=A_, **A_ ) def snake_case_ ( A_ : Any ): '''simple docstring''' with open(A_ ) as f: return json.load(A_ ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = git.Repo(search_parent_directories=A_ ) _lowerCamelCase : str = { '''repo_id''': str(A_ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case_ ( A_ : Callable, A_ : Iterable ): '''simple docstring''' return list(map(A_, A_ ) ) def snake_case_ ( A_ : str, A_ : Tuple ): '''simple docstring''' with open(A_, '''wb''' ) as f: return pickle.dump(A_, A_ ) def snake_case_ ( A_ : List[str] ): '''simple docstring''' def remove_articles(A_ : str ): return re.sub(R'''\b(a|an|the)\b''', ''' ''', A_ ) def white_space_fix(A_ : Any ): return " ".join(text.split() ) def remove_punc(A_ : List[Any] ): _lowerCamelCase : Any = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(A_ : Optional[int] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(A_ ) ) ) ) def snake_case_ ( A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : str = normalize_answer(A_ ).split() _lowerCamelCase : int = normalize_answer(A_ ).split() _lowerCamelCase : str = Counter(A_ ) & Counter(A_ ) _lowerCamelCase : Any = sum(common.values() ) if num_same == 0: return 0 _lowerCamelCase : int = 1.0 * num_same / len(A_ ) _lowerCamelCase : str = 1.0 * num_same / len(A_ ) _lowerCamelCase : List[Any] = (2 * precision * recall) / (precision + recall) return fa def snake_case_ ( A_ : Dict, A_ : str ): '''simple docstring''' return normalize_answer(A_ ) == normalize_answer(A_ ) def snake_case_ ( A_ : List[str], A_ : List[str] ): '''simple docstring''' assert len(A_ ) == len(A_ ) _lowerCamelCase : Optional[Any] = 0 for hypo, pred in zip(A_, A_ ): em += exact_match_score(A_, A_ ) if len(A_ ) > 0: em /= len(A_ ) return {"em": em} def snake_case_ ( A_ : Optional[int] ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def snake_case_ ( A_ : Dict, A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : Dict = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead _lowerCamelCase : Tuple = '''dropout_rate''' for p in extra_params: if getattr(A_, A_, A_ ): if not hasattr(A_, A_ ) and not hasattr(A_, equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(A_ ) ) delattr(A_, A_ ) continue _lowerCamelCase : Union[str, Any] = p if hasattr(A_, A_ ) else equivalent_param[p] setattr(A_, A_, getattr(A_, A_ ) ) delattr(A_, A_ ) return hparams, config
83
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. lowercase = [[1, 2, 4], [1, 2, 3, 4]] lowercase = DisjunctiveConstraint(snake_case ) self.assertTrue(isinstance(dc.token_ids , snake_case ) ) with self.assertRaises(snake_case ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(snake_case ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def SCREAMING_SNAKE_CASE__ ( self ): # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). lowercase = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(snake_case ): DisjunctiveConstraint(snake_case ) # fails here def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [[1, 2, 3], [1, 2, 4]] lowercase = DisjunctiveConstraint(snake_case ) lowercase , lowercase , lowercase = dc.update(1 ) lowercase = stepped is True and completed is False and reset is False self.assertTrue(snake_case ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) lowercase , lowercase , lowercase = dc.update(2 ) lowercase = stepped is True and completed is False and reset is False self.assertTrue(snake_case ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) lowercase , lowercase , lowercase = dc.update(3 ) lowercase = stepped is True and completed is True and reset is False self.assertTrue(snake_case ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] lowercase = DisjunctiveConstraint(snake_case ) lowercase , lowercase , lowercase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) lowercase , lowercase , lowercase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) lowercase , lowercase , lowercase = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) lowercase , lowercase , lowercase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() lowercase , lowercase , lowercase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) lowercase , lowercase , lowercase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) lowercase , lowercase , lowercase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
84
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
0
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): lowercase_ = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self : str , a_ : int , a_ : int , a_ : Optional[int] = None , a_ : int = 5_0257 , a_ : int = 1024 , a_ : int = 768 , a_ : int = 12 , a_ : int = 12 , a_ : Optional[int] = None , a_ : str = "gelu_new" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 1e-5 , a_ : float = 0.02 , a_ : bool = True , a_ : bool = True , a_ : bool = False , a_ : bool = False , )-> Optional[int]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ : List[Any] = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = prefix_inner_dim SCREAMING_SNAKE_CASE__ : str = prefix_hidden_dim SCREAMING_SNAKE_CASE__ : str = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) SCREAMING_SNAKE_CASE__ : List[Any] = ( nn.Linear(self.prefix_hidden_dim , a_ ) if self.prefix_hidden_dim is not None else nn.Identity() ) SCREAMING_SNAKE_CASE__ : List[str] = GPTaConfig( vocab_size=a_ , n_positions=a_ , n_embd=a_ , n_layer=a_ , n_head=a_ , n_inner=a_ , activation_function=a_ , resid_pdrop=a_ , embd_pdrop=a_ , attn_pdrop=a_ , layer_norm_epsilon=a_ , initializer_range=a_ , scale_attn_weights=a_ , use_cache=a_ , scale_attn_by_inverse_layer_idx=a_ , reorder_and_upcast_attn=a_ , ) SCREAMING_SNAKE_CASE__ : Optional[Any] = GPTaLMHeadModel(a_ ) def __lowercase( self : Any , a_ : torch.Tensor , a_ : torch.Tensor , a_ : Optional[torch.Tensor] = None , a_ : Optional[torch.Tensor] = None , )-> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.transformer.transformer.wte(a_ ) SCREAMING_SNAKE_CASE__ : List[str] = self.encode_prefix(a_ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = self.decode_prefix(a_ ) SCREAMING_SNAKE_CASE__ : Any = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) SCREAMING_SNAKE_CASE__ : Tuple = torch.cat((dummy_token, input_ids) , dim=1 ) SCREAMING_SNAKE_CASE__ : Any = self.transformer(inputs_embeds=a_ , labels=a_ , attention_mask=a_ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __lowercase( self : str , a_ : int , a_ : torch.device )-> torch.Tensor: """simple docstring""" return torch.zeros(a_ , self.prefix_length , dtype=torch.intaa , device=a_ ) def __lowercase( self : str , a_ : int )-> Any: """simple docstring""" return self.encode_prefix(a_ ) @torch.no_grad() def __lowercase( self : List[Any] , a_ : Tuple , a_ : int , a_ : str )-> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = torch.split(a_ , 1 , dim=0 ) SCREAMING_SNAKE_CASE__ : int = [] SCREAMING_SNAKE_CASE__ : Dict = [] for feature in features: SCREAMING_SNAKE_CASE__ : Dict = self.decode_prefix(feature.to(a_ ) ) # back to the clip feature # Only support beam search for now SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] = self.generate_beam( input_embeds=a_ , device=a_ , eos_token_id=a_ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) SCREAMING_SNAKE_CASE__ : Optional[int] = torch.stack(a_ ) SCREAMING_SNAKE_CASE__ : Optional[int] = torch.stack(a_ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __lowercase( self : Tuple , a_ : Optional[int]=None , a_ : Optional[Any]=None , a_ : str=None , a_ : int = 5 , a_ : int = 67 , a_ : float = 1.0 , a_ : Optional[int] = None , )-> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = eos_token_id SCREAMING_SNAKE_CASE__ : Tuple = None SCREAMING_SNAKE_CASE__ : Optional[int] = None SCREAMING_SNAKE_CASE__ : List[str] = torch.ones(a_ , device=a_ , dtype=torch.int ) SCREAMING_SNAKE_CASE__ : List[str] = torch.zeros(a_ , device=a_ , dtype=torch.bool ) if input_embeds is not None: SCREAMING_SNAKE_CASE__ : Dict = input_embeds else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.transformer.transformer.wte(a_ ) for i in range(a_ ): SCREAMING_SNAKE_CASE__ : Dict = self.transformer(inputs_embeds=a_ ) SCREAMING_SNAKE_CASE__ : int = outputs.logits SCREAMING_SNAKE_CASE__ : Optional[int] = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) SCREAMING_SNAKE_CASE__ : Tuple = logits.softmax(-1 ).log() if scores is None: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = logits.topk(a_ , -1 ) SCREAMING_SNAKE_CASE__ : List[str] = generated.expand(a_ , *generated.shape[1:] ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: SCREAMING_SNAKE_CASE__ : Dict = next_tokens else: SCREAMING_SNAKE_CASE__ : Dict = tokens.expand(a_ , *tokens.shape[1:] ) SCREAMING_SNAKE_CASE__ : List[Any] = torch.cat((tokens, next_tokens) , dim=1 ) else: SCREAMING_SNAKE_CASE__ : List[str] = -float(np.inf ) SCREAMING_SNAKE_CASE__ : Any = 0 SCREAMING_SNAKE_CASE__ : Any = scores[:, None] + logits seq_lengths[~is_stopped] += 1 SCREAMING_SNAKE_CASE__ : Dict = scores_sum / seq_lengths[:, None] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = scores_sum_average.view(-1 ).topk(a_ , -1 ) SCREAMING_SNAKE_CASE__ : str = next_tokens // scores_sum.shape[1] SCREAMING_SNAKE_CASE__ : str = seq_lengths[next_tokens_source] SCREAMING_SNAKE_CASE__ : Union[str, Any] = next_tokens % scores_sum.shape[1] SCREAMING_SNAKE_CASE__ : Any = next_tokens.unsqueeze(1 ) SCREAMING_SNAKE_CASE__ : int = tokens[next_tokens_source] SCREAMING_SNAKE_CASE__ : List[str] = torch.cat((tokens, next_tokens) , dim=1 ) SCREAMING_SNAKE_CASE__ : Any = generated[next_tokens_source] SCREAMING_SNAKE_CASE__ : List[str] = scores_sum_average * seq_lengths SCREAMING_SNAKE_CASE__ : int = is_stopped[next_tokens_source] SCREAMING_SNAKE_CASE__ : int = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) SCREAMING_SNAKE_CASE__ : List[Any] = torch.cat((generated, next_token_embed) , dim=1 ) SCREAMING_SNAKE_CASE__ : Tuple = is_stopped + next_tokens.eq(a_ ).squeeze() if is_stopped.all(): break SCREAMING_SNAKE_CASE__ : Dict = scores / seq_lengths SCREAMING_SNAKE_CASE__ : List[Any] = scores.argsort(descending=a_ ) # tokens tensors are already padded to max_seq_length SCREAMING_SNAKE_CASE__ : Optional[Any] = [tokens[i] for i in order] SCREAMING_SNAKE_CASE__ : Any = torch.stack(a_ , dim=0 ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
85
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __a :Optional[Any] = logging.get_logger(__name__) __a :Dict = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __a :Dict = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 1_0563, 1_0786, 1_1420, 1_1709, 1_1907, 1_3163, 1_3697, 1_3700, 1_4808, 1_5306, 1_6410, 1_6791, 1_7992, 1_9203, 1_9510, 2_0724, 2_2305, 2_2935, 2_7007, 3_0109, 3_0420, 3_3409, 3_4949, 4_0283, 4_0493, 4_0549, 4_7282, 4_9146, 5_0257, 5_0359, 5_0360, 5_0361 ] __a :Optional[int] = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 1_0428, 1_0929, 1_1938, 1_2033, 1_2331, 1_2562, 1_3793, 1_4157, 1_4635, 1_5265, 1_5618, 1_6553, 1_6604, 1_8362, 1_8956, 2_0075, 2_1675, 2_2520, 2_6130, 2_6161, 2_6435, 2_8279, 2_9464, 3_1650, 3_2302, 3_2470, 3_6865, 4_2863, 4_7425, 4_9870, 5_0254, 5_0258, 5_0360, 5_0361, 5_0362 ] class _a ( snake_case_ ): """simple docstring""" _lowerCamelCase : Tuple = 'whisper' _lowerCamelCase : Optional[Any] = ['past_key_values'] _lowerCamelCase : List[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self : Union[str, Any] , UpperCAmelCase : Tuple=51865 , UpperCAmelCase : Dict=80 , UpperCAmelCase : Any=6 , UpperCAmelCase : Optional[Any]=4 , UpperCAmelCase : Optional[int]=6 , UpperCAmelCase : Any=4 , UpperCAmelCase : str=1536 , UpperCAmelCase : Optional[int]=1536 , UpperCAmelCase : Optional[Any]=0.0 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : Tuple=50257 , UpperCAmelCase : List[str]=True , UpperCAmelCase : str=True , UpperCAmelCase : str="gelu" , UpperCAmelCase : Optional[Any]=256 , UpperCAmelCase : Union[str, Any]=0.0 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : str=0.0 , UpperCAmelCase : List[str]=0.02 , UpperCAmelCase : List[str]=False , UpperCAmelCase : str=1500 , UpperCAmelCase : Any=448 , UpperCAmelCase : str=50256 , UpperCAmelCase : List[Any]=50256 , UpperCAmelCase : Any=50256 , UpperCAmelCase : int=None , UpperCAmelCase : Optional[Any]=[220, 50256] , UpperCAmelCase : Optional[Any]=False , UpperCAmelCase : int=256 , UpperCAmelCase : List[str]=False , UpperCAmelCase : Dict=0.05 , UpperCAmelCase : List[str]=10 , UpperCAmelCase : int=2 , UpperCAmelCase : Tuple=0.0 , UpperCAmelCase : int=10 , UpperCAmelCase : Optional[int]=0 , UpperCAmelCase : Dict=7 , **UpperCAmelCase : List[Any] , ): A_ = vocab_size A_ = num_mel_bins A_ = d_model A_ = encoder_layers A_ = encoder_attention_heads A_ = decoder_layers A_ = decoder_attention_heads A_ = decoder_ffn_dim A_ = encoder_ffn_dim A_ = dropout A_ = attention_dropout A_ = activation_dropout A_ = activation_function A_ = init_std A_ = encoder_layerdrop A_ = decoder_layerdrop A_ = use_cache A_ = encoder_layers A_ = scale_embedding # scale factor will be sqrt(d_model) if True A_ = max_source_positions A_ = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. A_ = classifier_proj_size A_ = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 A_ = apply_spec_augment A_ = mask_time_prob A_ = mask_time_length A_ = mask_time_min_masks A_ = mask_feature_prob A_ = mask_feature_length A_ = mask_feature_min_masks A_ = median_filter_width super().__init__( pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , is_encoder_decoder=UpperCAmelCase , decoder_start_token_id=UpperCAmelCase , suppress_tokens=UpperCAmelCase , begin_suppress_tokens=UpperCAmelCase , **UpperCAmelCase , ) class _a ( snake_case_ ): """simple docstring""" @property def __A ( self : Optional[Any] ): A_ = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ] ) if self.use_past: A_ = {0: "batch"} else: A_ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(UpperCAmelCase , direction="inputs" ) return common_inputs def __A ( self : List[Any] , UpperCAmelCase : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional["TensorType"] = None , UpperCAmelCase : int = 22050 , UpperCAmelCase : float = 5.0 , UpperCAmelCase : int = 220 , ): A_ = OrderedDict() A_ = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=UpperCAmelCase , framework=UpperCAmelCase , sampling_rate=UpperCAmelCase , time_duration=UpperCAmelCase , frequency=UpperCAmelCase , ) A_ = encoder_inputs["input_features"].shape[2] A_ = encoder_sequence_length // 2 if self.use_past else seq_length A_ = super().generate_dummy_inputs( preprocessor.tokenizer , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) A_ = encoder_inputs.pop("input_features" ) A_ = decoder_inputs.pop("decoder_input_ids" ) if "past_key_values" in decoder_inputs: A_ = decoder_inputs.pop("past_key_values" ) return dummy_inputs @property def __A ( self : Dict ): return 1E-3
86
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ , ) -> tuple[str, float]: """simple docstring""" if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif stress < 0: raise ValueError('''Stress cannot be negative''' ) elif tangential_force < 0: raise ValueError('''Tangential Force cannot be negative''' ) elif area < 0: raise ValueError('''Area cannot be negative''' ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
87
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
0
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def _snake_case ( ): """simple docstring""" _lowerCamelCase : Optional[Any] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) _lowerCamelCase : Dict = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__snake_case ) # Let's go _lowerCamelCase : Any = parser.parse_args() if not hasattr(__snake_case , """func""" ): parser.print_help() exit(1 ) # Run _lowerCamelCase : List[str] = args.func(__snake_case ) service.run() if __name__ == "__main__": main()
88
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_ernie": ["ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ErnieConfig", "ErnieOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Optional[Any] = [ "ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST", "ErnieForCausalLM", "ErnieForMaskedLM", "ErnieForMultipleChoice", "ErnieForNextSentencePrediction", "ErnieForPreTraining", "ErnieForQuestionAnswering", "ErnieForSequenceClassification", "ErnieForTokenClassification", "ErnieModel", "ErniePreTrainedModel", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
0
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __UpperCAmelCase = logging.get_logger(__name__) # General docstring __UpperCAmelCase = '''MobileNetV1Config''' # Base docstring __UpperCAmelCase = '''google/mobilenet_v1_1.0_224''' __UpperCAmelCase = [1, 1_024, 7, 7] # Image classification docstring __UpperCAmelCase = '''google/mobilenet_v1_1.0_224''' __UpperCAmelCase = '''tabby, tabby cat''' __UpperCAmelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def _snake_case ( A , A , A=None ) -> Optional[int]: lowerCAmelCase__ = {} if isinstance(A , A ): lowerCAmelCase__ = model.mobilenet_va else: lowerCAmelCase__ = model lowerCAmelCase__ = '''MobilenetV1/Conv2d_0/''' lowerCAmelCase__ = backbone.conv_stem.convolution.weight lowerCAmelCase__ = backbone.conv_stem.normalization.bias lowerCAmelCase__ = backbone.conv_stem.normalization.weight lowerCAmelCase__ = backbone.conv_stem.normalization.running_mean lowerCAmelCase__ = backbone.conv_stem.normalization.running_var for i in range(13 ): lowerCAmelCase__ = i + 1 lowerCAmelCase__ = i * 2 lowerCAmelCase__ = backbone.layer[pt_index] lowerCAmelCase__ = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" lowerCAmelCase__ = pointer.convolution.weight lowerCAmelCase__ = pointer.normalization.bias lowerCAmelCase__ = pointer.normalization.weight lowerCAmelCase__ = pointer.normalization.running_mean lowerCAmelCase__ = pointer.normalization.running_var lowerCAmelCase__ = backbone.layer[pt_index + 1] lowerCAmelCase__ = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" lowerCAmelCase__ = pointer.convolution.weight lowerCAmelCase__ = pointer.normalization.bias lowerCAmelCase__ = pointer.normalization.weight lowerCAmelCase__ = pointer.normalization.running_mean lowerCAmelCase__ = pointer.normalization.running_var if isinstance(A , A ): lowerCAmelCase__ = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowerCAmelCase__ = model.classifier.weight lowerCAmelCase__ = model.classifier.bias return tf_to_pt_map def _snake_case ( A , A , A ) -> Any: try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowerCAmelCase__ = tf.train.list_variables(A ) lowerCAmelCase__ = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) lowerCAmelCase__ = tf.train.load_variable(A , A ) lowerCAmelCase__ = array # Build TF to PyTorch weights loading map lowerCAmelCase__ = _build_tf_to_pytorch_map(A , A , A ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue lowerCAmelCase__ = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowerCAmelCase__ = np.transpose(A , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowerCAmelCase__ = array.squeeze().transpose() else: lowerCAmelCase__ = np.transpose(A , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) lowerCAmelCase__ = torch.from_numpy(A ) tf_weights.pop(A , A ) tf_weights.pop(name + '''/RMSProp''' , A ) tf_weights.pop(name + '''/RMSProp_1''' , A ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , A ) logger.info(F"""Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}""" ) return model def _snake_case ( A , A ) -> torch.Tensor: lowerCAmelCase__ , lowerCAmelCase__ = features.shape[-2:] lowerCAmelCase__ , lowerCAmelCase__ = conv_layer.stride lowerCAmelCase__ , lowerCAmelCase__ = conv_layer.kernel_size if in_height % stride_height == 0: lowerCAmelCase__ = max(kernel_height - stride_height , 0 ) else: lowerCAmelCase__ = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowerCAmelCase__ = max(kernel_width - stride_width , 0 ) else: lowerCAmelCase__ = max(kernel_width - (in_width % stride_width) , 0 ) lowerCAmelCase__ = pad_along_width // 2 lowerCAmelCase__ = pad_along_width - pad_left lowerCAmelCase__ = pad_along_height // 2 lowerCAmelCase__ = pad_along_height - pad_top lowerCAmelCase__ = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(A , A , '''constant''' , 0.0 ) class a__ ( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = 1 , lowerCamelCase_ = 1 , lowerCamelCase_ = False , lowerCamelCase_ = True , lowerCamelCase_ = True , ) -> None: super().__init__() lowerCAmelCase__ = config if in_channels % groups != 0: raise ValueError(F"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(F"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) lowerCAmelCase__ = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowerCAmelCase__ = nn.Convad( in_channels=lowerCamelCase_ , out_channels=lowerCamelCase_ , kernel_size=lowerCamelCase_ , stride=lowerCamelCase_ , padding=lowerCamelCase_ , groups=lowerCamelCase_ , bias=lowerCamelCase_ , padding_mode='''zeros''' , ) if use_normalization: lowerCAmelCase__ = nn.BatchNormad( num_features=lowerCamelCase_ , eps=config.layer_norm_eps , momentum=0.9_997 , affine=lowerCamelCase_ , track_running_stats=lowerCamelCase_ , ) else: lowerCAmelCase__ = None if use_activation: if isinstance(lowerCamelCase_ , lowerCamelCase_ ): lowerCAmelCase__ = ACTaFN[use_activation] elif isinstance(config.hidden_act , lowerCamelCase_ ): lowerCAmelCase__ = ACTaFN[config.hidden_act] else: lowerCAmelCase__ = config.hidden_act else: lowerCAmelCase__ = None def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> torch.Tensor: if self.config.tf_padding: lowerCAmelCase__ = apply_tf_padding(lowerCamelCase_ , self.convolution ) lowerCAmelCase__ = self.convolution(lowerCamelCase_ ) if self.normalization is not None: lowerCAmelCase__ = self.normalization(lowerCamelCase_ ) if self.activation is not None: lowerCAmelCase__ = self.activation(lowerCamelCase_ ) return features class a__ ( a__ ): '''simple docstring''' lowercase__ : int = MobileNetVaConfig lowercase__ : Optional[Any] = load_tf_weights_in_mobilenet_va lowercase__ : str = "mobilenet_v1" lowercase__ : Dict = "pixel_values" lowercase__ : Union[str, Any] = False def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> None: if isinstance(lowerCamelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(lowerCamelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __UpperCAmelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __UpperCAmelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , a__ , ) class a__ ( a__ ): '''simple docstring''' def __init__( self , lowerCamelCase_ , lowerCamelCase_ = True ) -> Dict: super().__init__(lowerCamelCase_ ) lowerCAmelCase__ = config lowerCAmelCase__ = 32 lowerCAmelCase__ = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowerCAmelCase__ = MobileNetVaConvLayer( lowerCamelCase_ , in_channels=config.num_channels , out_channels=lowerCamelCase_ , kernel_size=3 , stride=2 , ) lowerCAmelCase__ = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowerCAmelCase__ = nn.ModuleList() for i in range(13 ): lowerCAmelCase__ = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowerCAmelCase__ = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( lowerCamelCase_ , in_channels=lowerCamelCase_ , out_channels=lowerCamelCase_ , kernel_size=3 , stride=strides[i] , groups=lowerCamelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( lowerCamelCase_ , in_channels=lowerCamelCase_ , out_channels=lowerCamelCase_ , kernel_size=1 , ) ) lowerCAmelCase__ = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> Any: raise NotImplementedError @add_start_docstrings_to_model_forward(lowerCamelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase_ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowerCAmelCase__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowerCAmelCase__ = self.conv_stem(lowerCamelCase_ ) lowerCAmelCase__ = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowerCAmelCase__ = layer_module(lowerCamelCase_ ) if output_hidden_states: lowerCAmelCase__ = all_hidden_states + (hidden_states,) lowerCAmelCase__ = hidden_states if self.pooler is not None: lowerCAmelCase__ = torch.flatten(self.pooler(lowerCamelCase_ ) , start_dim=1 ) else: lowerCAmelCase__ = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase_ , pooler_output=lowerCamelCase_ , hidden_states=lowerCamelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , a__ , ) class a__ ( a__ ): '''simple docstring''' def __init__( self , lowerCamelCase_ ) -> None: super().__init__(lowerCamelCase_ ) lowerCAmelCase__ = config.num_labels lowerCAmelCase__ = MobileNetVaModel(lowerCamelCase_ ) lowerCAmelCase__ = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowerCAmelCase__ = nn.Dropout(config.classifier_dropout_prob , inplace=lowerCamelCase_ ) lowerCAmelCase__ = nn.Linear(lowerCamelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCamelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowerCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase__ = self.mobilenet_va(lowerCamelCase_ , output_hidden_states=lowerCamelCase_ , return_dict=lowerCamelCase_ ) lowerCAmelCase__ = outputs.pooler_output if return_dict else outputs[1] lowerCAmelCase__ = self.classifier(self.dropout(lowerCamelCase_ ) ) lowerCAmelCase__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCAmelCase__ = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCAmelCase__ = '''single_label_classification''' else: lowerCAmelCase__ = '''multi_label_classification''' if self.config.problem_type == "regression": lowerCAmelCase__ = MSELoss() if self.num_labels == 1: lowerCAmelCase__ = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowerCAmelCase__ = loss_fct(lowerCamelCase_ , lowerCamelCase_ ) elif self.config.problem_type == "single_label_classification": lowerCAmelCase__ = CrossEntropyLoss() lowerCAmelCase__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowerCAmelCase__ = BCEWithLogitsLoss() lowerCAmelCase__ = loss_fct(lowerCamelCase_ , lowerCamelCase_ ) if not return_dict: lowerCAmelCase__ = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=lowerCamelCase_ , logits=lowerCamelCase_ , hidden_states=outputs.hidden_states , )
90
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
0
"""simple docstring""" _lowercase = '''Tobias Carryer''' from time import time class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[Any] ,A_ : str ,A_ : Optional[Any] ,A_ : List[str] ,A_ : List[Any]=int(time() ) ) -> Union[str, Any]: # noqa: B008 A = multiplier A = increment A = modulo A = seed def _SCREAMING_SNAKE_CASE ( self : Any ) -> int: A = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. _lowercase = LinearCongruentialGenerator(1_66_45_25, 10_13_90_42_23, 2 << 31) while True: print(lcg.next_number())
91
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available UpperCamelCase_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ["""GPTSw3Tokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
92
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
0
"""simple docstring""" import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) def __A (_SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" lowerCAmelCase__ :List[Any] = torch.load(_SCREAMING_SNAKE_CASE , map_location='cpu' ) if "model" in sd.keys(): lowerCAmelCase__ :Optional[int] = torch.load(_SCREAMING_SNAKE_CASE , map_location='cpu' )['model'] # pop unnecessary weights lowerCAmelCase__ :Tuple = [ 'decoder.version', 'decoder.output_projection.weight', ] for key in keys_to_delete: if key in sd: sd.pop(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[Any] = { 'decoder.project_in_dim.weight': 'decoder.project_in.weight', 'decoder.project_out_dim.weight': 'decoder.project_out.weight', 'decoder.layer_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.layer_norm.bias': 'decoder.final_layer_norm.bias', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: lowerCAmelCase__ :Union[str, Any] = sd.pop(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Union[str, Any] = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: lowerCAmelCase__ :List[Any] = sd[key] # We split QKV in separate Q,K,V lowerCAmelCase__ :Any = key.replace('.qkv_proj.' , '.q_proj.' ) lowerCAmelCase__ :Optional[Any] = key.replace('.qkv_proj.' , '.k_proj.' ) lowerCAmelCase__ :List[str] = key.replace('.qkv_proj.' , '.v_proj.' ) lowerCAmelCase__ :Optional[int] = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :List[str] = torch.split(_SCREAMING_SNAKE_CASE , depth // 3 , dim=0 ) lowerCAmelCase__ :Optional[int] = q lowerCAmelCase__ :Optional[Any] = k lowerCAmelCase__ :Union[str, Any] = v del sd[key] return sd @torch.no_grad() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ :Dict = load_checkpoint(_SCREAMING_SNAKE_CASE ) if config is not None: lowerCAmelCase__ :str = OPTConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: lowerCAmelCase__ :List[str] = OPTConfig() lowerCAmelCase__ :Tuple = OPTModel(_SCREAMING_SNAKE_CASE ).half().eval() model.load_state_dict(_SCREAMING_SNAKE_CASE ) # Check results Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--fairseq_path""", type=str, help=( """path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:""" """ https://huggingface.co/models?other=opt_metasq""" ), ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--hf_config""", default=None, type=str, help="""Define HF config.""") __A = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
93
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 SCREAMING_SNAKE_CASE = data_utils.TransfoXLTokenizer SCREAMING_SNAKE_CASE = data_utils.TransfoXLCorpus SCREAMING_SNAKE_CASE = data_utils SCREAMING_SNAKE_CASE = data_utils def lowercase_ ( __A : int , __A : Any , __A : List[Any] , __A : Optional[Any] ) -> Dict: """simple docstring""" if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(__A , '''rb''' ) as fp: lowercase : Optional[int] =pickle.load(__A , encoding='''latin1''' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) lowercase : Any =pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''pretrained_vocab_file'''] print(F'Save vocabulary to {pytorch_vocab_dump_path}' ) lowercase : Optional[Any] =corpus.vocab.__dict__ torch.save(__A , __A ) lowercase : Optional[int] =corpus.__dict__ corpus_dict_no_vocab.pop('''vocab''' , __A ) lowercase : Dict =pytorch_dump_folder_path + '''/''' + CORPUS_NAME print(F'Save dataset to {pytorch_dataset_dump_path}' ) torch.save(__A , __A ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model lowercase : Any =os.path.abspath(__A ) lowercase : int =os.path.abspath(__A ) print(F'Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.' ) # Initialise PyTorch model if transfo_xl_config_file == "": lowercase : Tuple =TransfoXLConfig() else: lowercase : List[str] =TransfoXLConfig.from_json_file(__A ) print(F'Building PyTorch model from configuration: {config}' ) lowercase : List[Any] =TransfoXLLMHeadModel(__A ) lowercase : str =load_tf_weights_in_transfo_xl(__A , __A , __A ) # Save pytorch-model lowercase : str =os.path.join(__A , __A ) lowercase : Any =os.path.join(__A , __A ) print(F'Save PyTorch model to {os.path.abspath(__A )}' ) torch.save(model.state_dict() , __A ) print(F'Save configuration file to {os.path.abspath(__A )}' ) with open(__A , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
94
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
0
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class UpperCamelCase_ (__A ): def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: UpperCAmelCase_ : Dict = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , "num_attention_heads" ) ) class UpperCamelCase_ : def __init__( self : List[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : List[Any]=13 , lowerCAmelCase_ : Tuple=64 , lowerCAmelCase_ : str=3 , lowerCAmelCase_ : Optional[int]=3 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Optional[Any]=1 , lowerCAmelCase_ : Optional[int]=16 , lowerCAmelCase_ : Any=[128, 256, 384] , lowerCAmelCase_ : int=[4, 6, 8] , lowerCAmelCase_ : Optional[int]=[2, 3, 4] , lowerCAmelCase_ : List[Any]=[16, 16, 16] , lowerCAmelCase_ : List[str]=0 , lowerCAmelCase_ : int=[2, 2, 2] , lowerCAmelCase_ : List[Any]=[2, 2, 2] , lowerCAmelCase_ : str=0.0_2 , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : int=2 , ) -> Any: UpperCAmelCase_ : str = parent UpperCAmelCase_ : str = batch_size UpperCAmelCase_ : Tuple = image_size UpperCAmelCase_ : Dict = num_channels UpperCAmelCase_ : List[Any] = kernel_size UpperCAmelCase_ : List[Any] = stride UpperCAmelCase_ : Optional[int] = padding UpperCAmelCase_ : Dict = hidden_sizes UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : Optional[int] = depths UpperCAmelCase_ : Union[str, Any] = key_dim UpperCAmelCase_ : Any = drop_path_rate UpperCAmelCase_ : List[str] = patch_size UpperCAmelCase_ : str = attention_ratio UpperCAmelCase_ : str = mlp_ratio UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : Union[str, Any] = [ ["Subsample", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["Subsample", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] UpperCAmelCase_ : List[Any] = is_training UpperCAmelCase_ : Optional[Any] = use_labels UpperCAmelCase_ : Dict = num_labels UpperCAmelCase_ : Any = initializer_range def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Tuple: UpperCAmelCase_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : List[str] = None if self.use_labels: UpperCAmelCase_ : Dict = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ : Optional[Any] = self.get_config() return config, pixel_values, labels def _SCREAMING_SNAKE_CASE ( self : int ) -> Optional[Any]: return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[Any] ) -> Optional[int]: UpperCAmelCase_ : Any = LevitModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Any = model(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = (self.image_size, self.image_size) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = image_size[0], image_size[1] for _ in range(4 ): UpperCAmelCase_ : List[str] = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) UpperCAmelCase_ : Dict = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : int ) -> Optional[Any]: UpperCAmelCase_ : Optional[int] = self.num_labels UpperCAmelCase_ : Optional[Any] = LevitForImageClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() UpperCAmelCase_ : Dict = model(lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: UpperCAmelCase_ : str = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : int = config_and_inputs UpperCAmelCase_ : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase_ (__A , __A , unittest.TestCase ): __magic_name__ = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) __magic_name__ = ( { '''feature-extraction''': LevitModel, '''image-classification''': (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def _SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: UpperCAmelCase_ : str = LevitModelTester(self ) UpperCAmelCase_ : Any = ConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ , hidden_size=37 ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _SCREAMING_SNAKE_CASE ( self : int ) -> Tuple: return @unittest.skip(reason="Levit does not use inputs_embeds" ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> str: pass @unittest.skip(reason="Levit does not support input and output embeddings" ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[int]: pass @unittest.skip(reason="Levit does not output attentions" ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: pass def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Any = model_class(lowerCAmelCase_ ) UpperCAmelCase_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : int = [*signature.parameters.keys()] UpperCAmelCase_ : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> Tuple: def check_hidden_states_output(lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[str] ): UpperCAmelCase_ : Dict = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Any = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) ) UpperCAmelCase_ : str = outputs.hidden_states UpperCAmelCase_ : List[str] = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowerCAmelCase_ ) , lowerCAmelCase_ ) UpperCAmelCase_ : Dict = (self.model_tester.image_size, self.model_tester.image_size) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = image_size[0], image_size[1] for _ in range(4 ): UpperCAmelCase_ : Dict = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) UpperCAmelCase_ : List[Any] = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) UpperCAmelCase_ , UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : List[str] = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : str = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[str]: pass def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Dict=False ) -> Optional[Any]: UpperCAmelCase_ : List[Any] = super()._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _SCREAMING_SNAKE_CASE ( self : Any ) -> Any: UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[Any]: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: if not self.model_tester.is_training: return UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Optional[Any] = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCAmelCase_ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue UpperCAmelCase_ : int = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.train() UpperCAmelCase_ : Dict = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) UpperCAmelCase_ : Any = model(**lowerCAmelCase_ ).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase_ : Tuple = False UpperCAmelCase_ : Tuple = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase_ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue UpperCAmelCase_ : List[Any] = model_class(lowerCAmelCase_ ) model.gradient_checkpointing_enable() model.to(lowerCAmelCase_ ) model.train() UpperCAmelCase_ : Optional[int] = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) UpperCAmelCase_ : Any = model(**lowerCAmelCase_ ).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Any: UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Tuple = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCAmelCase_ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): UpperCAmelCase_ : int = problem_type["title"] UpperCAmelCase_ : int = problem_type["num_labels"] UpperCAmelCase_ : Dict = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.train() UpperCAmelCase_ : int = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) if problem_type["num_labels"] > 1: UpperCAmelCase_ : Optional[Any] = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) UpperCAmelCase_ : str = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCAmelCase_ ) as warning_list: UpperCAmelCase_ : Dict = model(**lowerCAmelCase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> str: for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[int] = LevitModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def snake_case ( ): UpperCAmelCase_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCamelCase_ (unittest.TestCase ): @cached_property def _SCREAMING_SNAKE_CASE ( self : Any ) -> Optional[Any]: return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def _SCREAMING_SNAKE_CASE ( self : str ) -> str: UpperCAmelCase_ : Optional[Any] = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowerCAmelCase_ ) UpperCAmelCase_ : Any = self.default_image_processor UpperCAmelCase_ : Dict = prepare_img() UpperCAmelCase_ : int = image_processor(images=lowerCAmelCase_ , return_tensors="pt" ).to(lowerCAmelCase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(**lowerCAmelCase_ ) # verify the logits UpperCAmelCase_ : Dict = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) UpperCAmelCase_ : Dict = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase_ , atol=1e-4 ) )
95
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
0
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, RobertaTokenizer, RobertaTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ ,unittest.TestCase ): UpperCAmelCase__ = RobertaTokenizer UpperCAmelCase__ = RobertaTokenizerFast UpperCAmelCase__ = True UpperCAmelCase__ = {"cls_token": "<s>"} def lowerCamelCase__ ( self : List[str] ) -> Dict: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __magic_name__: List[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] __magic_name__: List[str] = dict(zip(__snake_case , range(len(__snake_case ) ) ) ) __magic_name__: Tuple = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __magic_name__: Optional[int] = {"""unk_token""": """<unk>"""} __magic_name__: List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __magic_name__: int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__snake_case ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__snake_case ) ) def lowerCamelCase__ ( self : Optional[Any] , **__snake_case : str ) -> str: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__snake_case ) def lowerCamelCase__ ( self : Any , **__snake_case : Optional[Any] ) -> Optional[Any]: kwargs.update(self.special_tokens_map ) return RobertaTokenizerFast.from_pretrained(self.tmpdirname , **__snake_case ) def lowerCamelCase__ ( self : List[str] , __snake_case : Optional[Any] ) -> List[Any]: __magic_name__: List[str] = """lower newer""" __magic_name__: Optional[int] = """lower newer""" return input_text, output_text def lowerCamelCase__ ( self : List[str] ) -> Optional[Any]: __magic_name__: List[str] = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __magic_name__: List[Any] = """lower newer""" __magic_name__: List[str] = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] __magic_name__: Optional[int] = tokenizer.tokenize(__snake_case ) # , add_prefix_space=True) self.assertListEqual(__snake_case , __snake_case ) __magic_name__: int = tokens + [tokenizer.unk_token] __magic_name__: Tuple = [0, 1, 2, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) , __snake_case ) def lowerCamelCase__ ( self : List[str] ) -> Optional[int]: __magic_name__: int = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""" , add_special_tokens=__snake_case ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""" , add_special_tokens=__snake_case ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2] , ) @slow def lowerCamelCase__ ( self : Any ) -> List[str]: __magic_name__: Any = self.tokenizer_class.from_pretrained("""roberta-base""" ) __magic_name__: List[str] = tokenizer.encode("""sequence builders""" , add_special_tokens=__snake_case ) __magic_name__: Union[str, Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__snake_case ) __magic_name__: Optional[Any] = tokenizer.encode( """sequence builders""" , add_special_tokens=__snake_case , add_prefix_space=__snake_case ) __magic_name__: List[str] = tokenizer.encode( """sequence builders""" , """multi-sequence build""" , add_special_tokens=__snake_case , add_prefix_space=__snake_case ) __magic_name__: Union[str, Any] = tokenizer.build_inputs_with_special_tokens(__snake_case ) __magic_name__: Optional[Any] = tokenizer.build_inputs_with_special_tokens(__snake_case , __snake_case ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def lowerCamelCase__ ( self : int ) -> str: __magic_name__: int = self.get_tokenizer() __magic_name__: Tuple = """Encode this sequence.""" __magic_name__: List[str] = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments __magic_name__: Optional[Any] = tokenizer.encode(__snake_case , add_special_tokens=__snake_case , add_prefix_space=__snake_case ) __magic_name__: Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(__snake_case , __snake_case ) __magic_name__: Union[str, Any] = tokenizer.encode(__snake_case , add_special_tokens=__snake_case , add_prefix_space=__snake_case ) __magic_name__: Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(__snake_case , __snake_case ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) __magic_name__: Union[str, Any] = tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) __magic_name__: Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(__snake_case , __snake_case ) # Testing spaces after special tokens __magic_name__: List[str] = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case )} ) # mask token has a left space __magic_name__: int = tokenizer.convert_tokens_to_ids(__snake_case ) __magic_name__: int = """Encode <mask> sequence""" __magic_name__: List[Any] = """Encode <mask>sequence""" __magic_name__: Union[str, Any] = tokenizer.encode(__snake_case ) __magic_name__: Optional[Any] = encoded.index(__snake_case ) __magic_name__: Tuple = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(__snake_case , __snake_case ) __magic_name__: List[str] = tokenizer.encode(__snake_case ) __magic_name__: Any = encoded.index(__snake_case ) __magic_name__: str = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(__snake_case , __snake_case ) def lowerCamelCase__ ( self : str ) -> int: pass def lowerCamelCase__ ( self : str ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): __magic_name__: Tuple = self.rust_tokenizer_class.from_pretrained(__snake_case , **__snake_case ) __magic_name__: List[Any] = self.tokenizer_class.from_pretrained(__snake_case , **__snake_case ) __magic_name__: List[Any] = """A, <mask> AllenNLP sentence.""" __magic_name__: Optional[Any] = tokenizer_r.encode_plus(__snake_case , add_special_tokens=__snake_case , return_token_type_ids=__snake_case ) __magic_name__: str = tokenizer_p.encode_plus(__snake_case , add_special_tokens=__snake_case , return_token_type_ids=__snake_case ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) __magic_name__: List[str] = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) __magic_name__: List[Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual( __snake_case , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( __snake_case , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def lowerCamelCase__ ( self : Tuple ) -> str: for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __magic_name__: Union[str, Any] = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Any = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __magic_name__: Any = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""] , __snake_case ) self.assertEqual(post_processor_state["""add_prefix_space"""] , __snake_case ) self.assertEqual(post_processor_state["""trim_offsets"""] , __snake_case ) def lowerCamelCase__ ( self : Any ) -> Tuple: # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): __magic_name__: List[Any] = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` __magic_name__: Union[str, Any] = F'{text_of_1_token} {text_of_1_token}' __magic_name__: Dict = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Union[str, Any] = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__snake_case ) + 1, len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: Union[str, Any] = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: str = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__snake_case ) + 1, len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: List[str] = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Any = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__snake_case ), len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: int = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Dict = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__snake_case ), len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: Dict = F' {text}' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __magic_name__: str = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Any = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__snake_case ) + 1, 1 + len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: int = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: str = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__snake_case ), 1 + len(__snake_case ) + 1 + len(__snake_case )) , ) __magic_name__: List[Any] = self.rust_tokenizer_class.from_pretrained( __snake_case , use_fast=__snake_case , add_prefix_space=__snake_case , trim_offsets=__snake_case ) __magic_name__: Union[str, Any] = tokenizer_r(__snake_case , return_offsets_mapping=__snake_case , add_special_tokens=__snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__snake_case ), 1 + len(__snake_case ) + 1 + len(__snake_case )) , )
96
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
0
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor __a = logging.get_logger(__name__) class lowercase__( UpperCAmelCase ): """simple docstring""" def __init__( self : Dict , *SCREAMING_SNAKE_CASE_ : Tuple , **SCREAMING_SNAKE_CASE_ : Any ) -> None: warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , SCREAMING_SNAKE_CASE_ , ) super().__init__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ )
97
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
0
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def a__ ( lowercase : Tuple, lowercase : Optional[int], lowercase : str=None ) -> Tuple: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F"""{torch_layer} layer.weight does not match""" _UpperCamelCase = nn.Parameter(lowercase ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F"""{torch_layer} layer.bias does not match""" _UpperCamelCase = nn.Parameter(lowercase ) def a__ ( lowercase : Optional[int], lowercase : List[Any], lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.asarray(weights[0] ) _UpperCamelCase = np.asarray(weights[1] ) _UpperCamelCase = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key, torch.tensor(lowercase ).transpose(1, 2 ).contiguous().view(-1, lowercase ), ) set_param( torch_layer.self_attention.value, torch.tensor(lowercase ).transpose(1, 2 ).contiguous().view(-1, lowercase ), ) set_param( torch_layer.output.dense, torch.tensor(lowercase ).view(-1, lowercase ).contiguous().transpose(0, 1 ), ) def a__ ( lowercase : int, lowercase : int, lowercase : str ) -> Dict: """simple docstring""" _UpperCamelCase = np.asarray(weights[0] ) _UpperCamelCase = np.asarray(weights[1] ) _UpperCamelCase = np.asarray(weights[2] ) _UpperCamelCase = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query, torch.tensor(lowercase ).transpose(1, 2 ).contiguous().view(-1, lowercase ), ) set_param( torch_layer.self_attention.key, torch.tensor(lowercase ).transpose(1, 2 ).contiguous().view(-1, lowercase ), ) set_param( torch_layer.self_attention.value, torch.tensor(lowercase ).transpose(1, 2 ).contiguous().view(-1, lowercase ), ) set_param( torch_layer.output.dense, torch.tensor(lowercase ).view(-1, lowercase ).contiguous().transpose(0, 1 ), ) def a__ ( lowercase : List[Any], lowercase : List[str], lowercase : Tuple ) -> Dict: """simple docstring""" _UpperCamelCase = weights[0][0][0] _UpperCamelCase = np.asarray(layer_norm_a[0] ) _UpperCamelCase = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm, torch.tensor(lowercase ), torch.tensor(lowercase ), ) # lsh weights + output _UpperCamelCase = weights[0][1] if len(lowercase ) < 4: set_layer_weights_in_torch_lsh(lowercase, torch_block.attention, lowercase ) else: set_layer_weights_in_torch_local(lowercase, torch_block.attention, lowercase ) # intermediate weighs _UpperCamelCase = weights[2][0][1][2] # Chunked Feed Forward if len(lowercase ) == 4: _UpperCamelCase = intermediate_weights[2] # layernorm 2 _UpperCamelCase = np.asarray(intermediate_weights[0][0] ) _UpperCamelCase = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm, torch.tensor(lowercase ), torch.tensor(lowercase ), ) # intermediate dense _UpperCamelCase = np.asarray(intermediate_weights[1][0] ) _UpperCamelCase = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense, torch.tensor(lowercase ).transpose(0, 1 ).contiguous(), torch.tensor(lowercase ), ) # intermediate out _UpperCamelCase = np.asarray(intermediate_weights[4][0] ) _UpperCamelCase = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense, torch.tensor(lowercase ).transpose(0, 1 ).contiguous(), torch.tensor(lowercase ), ) def a__ ( lowercase : int, lowercase : str, lowercase : List[str] ) -> str: """simple docstring""" _UpperCamelCase = torch_model.reformer # word embeds _UpperCamelCase = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings, torch.tensor(lowercase ), ) if isinstance(weights[3], lowercase ): _UpperCamelCase = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _UpperCamelCase = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F"""{position_embeddings[emb_idx]} emb does not match""" _UpperCamelCase = nn.Parameter(torch.tensor(lowercase ) ) _UpperCamelCase = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( lowercase ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _UpperCamelCase = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(lowercase, lowercase, lowercase ) # output layer norm _UpperCamelCase = np.asarray(weights[7][0] ) _UpperCamelCase = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm, torch.tensor(lowercase ), torch.tensor(lowercase ), ) # output embeddings _UpperCamelCase = np.asarray(weights[9][0] ) _UpperCamelCase = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder, torch.tensor(lowercase ).transpose(0, 1 ).contiguous(), torch.tensor(lowercase ), ) def a__ ( lowercase : List[Any], lowercase : Optional[int], lowercase : Tuple ) -> List[Any]: """simple docstring""" _UpperCamelCase = ReformerConfig.from_json_file(lowercase ) print(F"""Building PyTorch model from configuration: {config}""" ) _UpperCamelCase = ReformerModelWithLMHead(lowercase ) with open(lowercase, '''rb''' ) as f: _UpperCamelCase = pickle.load(lowercase )['''weights'''] set_model_weights_in_torch(lowercase, lowercase, config.hidden_size ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict(), lowercase ) if __name__ == "__main__": lowercase__ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--trax_model_pkl_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained Reformer model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowercase__ : Union[str, Any] = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
98
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
from __future__ import annotations def a (lowerCAmelCase__ , lowerCAmelCase__ = None ): __a = word_bank or [] # create a table __a = len(lowerCAmelCase__ ) + 1 __a = [] for _ in range(lowerCAmelCase__ ): table.append([] ) # seed value __a = [[]] # because empty string has empty combination # iterate through the indices for i in range(lowerCAmelCase__ ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(lowerCAmelCase__ )] == word: __a = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(lowerCAmelCase__ )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(lowerCAmelCase__ )]: combination.reverse() return table[len(lowerCAmelCase__ )] if __name__ == "__main__": print(all_construct('jwajalapa', ['jwa', 'j', 'w', 'a', 'la', 'lapa'])) print(all_construct('rajamati', ['s', 'raj', 'amat', 'raja', 'ma', 'i', 't'])) print( all_construct( 'hexagonosaurus', ['h', 'ex', 'hex', 'ag', 'ago', 'ru', 'auru', 'rus', 'go', 'no', 'o', 's'], ) )
99
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
0
def __snake_case ( ) -> int: return 1 def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ ) -> int: return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ = 2_0_0 ) -> int: return two_pound(lowerCAmelCase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
100
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
0
from collections import defaultdict class __lowercase : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 SCREAMING_SNAKE_CASE_ : Any = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(lowerCAmelCase__ ) ) ] SCREAMING_SNAKE_CASE_ : Dict = defaultdict(lowerCAmelCase__ ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 SCREAMING_SNAKE_CASE_ : Optional[int] = (1 << len(lowerCAmelCase__ )) - 1 def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement SCREAMING_SNAKE_CASE_ : int = self.count_ways_until(lowerCAmelCase__ , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. SCREAMING_SNAKE_CASE_ : Optional[int] = total_ways_util return self.dp[mask][task_no] def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" for i in range(len(lowerCAmelCase__ ) ): for j in task_performed[i]: self.task[j].append(lowerCAmelCase__ ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": lowerCAmelCase__ : str =5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. lowerCAmelCase__ : List[Any] =[[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
101
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : str = logging.get_logger(__name__) __magic_name__ : Union[str, Any] = { """RWKV/rwkv-4-169m-pile""": """https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json""", """RWKV/rwkv-4-430m-pile""": """https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json""", """RWKV/rwkv-4-1b5-pile""": """https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json""", """RWKV/rwkv-4-3b-pile""": """https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json""", """RWKV/rwkv-4-7b-pile""": """https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json""", """RWKV/rwkv-4-14b-pile""": """https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json""", """RWKV/rwkv-raven-1b5""": """https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json""", """RWKV/rwkv-raven-3b""": """https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json""", """RWKV/rwkv-raven-7b""": """https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json""", """RWKV/rwkv-raven-14b""": """https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json""", } class lowercase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase : Any = """rwkv""" __lowerCAmelCase : List[str] = {"""max_position_embeddings""": """context_length"""} def __init__( self , _A=5_0_2_7_7 , _A=1_0_2_4 , _A=4_0_9_6 , _A=3_2 , _A=None , _A=None , _A=1e-5 , _A=0 , _A=0 , _A=6 , _A=False , _A=True , **_A , ): '''simple docstring''' UpperCamelCase : Any = vocab_size UpperCamelCase : Optional[Any] = context_length UpperCamelCase : str = hidden_size UpperCamelCase : int = num_hidden_layers UpperCamelCase : Dict = attention_hidden_size if attention_hidden_size is not None else hidden_size UpperCamelCase : Any = intermediate_size if intermediate_size is not None else 4 * hidden_size UpperCamelCase : Any = layer_norm_epsilon UpperCamelCase : Optional[int] = rescale_every UpperCamelCase : Optional[int] = use_cache UpperCamelCase : Union[str, Any] = bos_token_id UpperCamelCase : Any = eos_token_id super().__init__( tie_word_embeddings=_A , bos_token_id=_A , eos_token_id=_A , **_A )
102
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
0
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase : def __init__( self : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Any=1_3 , __lowerCamelCase : int=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : Any=3 , __lowerCamelCase : List[str]=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=3_2 , __lowerCamelCase : Optional[Any]=5 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Optional[Any]=3_7 , __lowerCamelCase : Optional[Any]="gelu" , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Tuple=1_0 , __lowerCamelCase : List[str]=0.0_2 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=2 , ): """simple docstring""" _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = patch_size _snake_case = num_channels _snake_case = is_training _snake_case = use_labels _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = scope _snake_case = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _snake_case = (image_size // patch_size) ** 2 _snake_case = num_patches + 1 def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : Tuple ): """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple ): """simple docstring""" _snake_case = ViTModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() _snake_case = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self : Any , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : str ): """simple docstring""" _snake_case = ViTForMaskedImageModeling(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() _snake_case = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _snake_case = 1 _snake_case = ViTForMaskedImageModeling(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() _snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _snake_case = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __UpperCAmelCase ( self : Dict , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple ): """simple docstring""" _snake_case = self.type_sequence_label_size _snake_case = ViTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() _snake_case = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _snake_case = 1 _snake_case = ViTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() _snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _snake_case = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self : Dict ): """simple docstring""" _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = config_and_inputs _snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( __SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,unittest.TestCase ): A__ : Optional[int] = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) A__ : Dict = ( {'''feature-extraction''': ViTModel, '''image-classification''': ViTForImageClassification} if is_torch_available() else {} ) A__ : Optional[int] = True A__ : str = False A__ : Any = False A__ : List[Any] = False def __UpperCAmelCase ( self : List[str] ): """simple docstring""" _snake_case = ViTModelTester(self ) _snake_case = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def __UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" pass def __UpperCAmelCase ( self : Dict ): """simple docstring""" _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , nn.Linear ) ) def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(__lowerCamelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def __UpperCAmelCase ( self : str ): """simple docstring""" _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def __UpperCAmelCase ( self : str ): """simple docstring""" _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = ViTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def snake_case ( ) -> Any: _snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def __UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" _snake_case = ViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ).to(__lowerCamelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=__lowerCamelCase , return_tensors='''pt''' ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): _snake_case = model(**__lowerCamelCase ) # verify the logits _snake_case = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) _snake_case = torch.tensor([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) ) @slow def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. _snake_case = ViTModel.from_pretrained('''facebook/dino-vits8''' ).to(__lowerCamelCase ) _snake_case = ViTImageProcessor.from_pretrained('''facebook/dino-vits8''' , size=4_8_0 ) _snake_case = prepare_img() _snake_case = image_processor(images=__lowerCamelCase , return_tensors='''pt''' ) _snake_case = inputs.pixel_values.to(__lowerCamelCase ) # forward pass with torch.no_grad(): _snake_case = model(__lowerCamelCase , interpolate_pos_encoding=__lowerCamelCase ) # verify the logits _snake_case = torch.Size((1, 3_6_0_1, 3_8_4) ) self.assertEqual(outputs.last_hidden_state.shape , __lowerCamelCase ) _snake_case = torch.tensor( [[4.2_3_4_0, 4.3_9_0_6, -6.6_6_9_2], [4.5_4_6_3, 1.8_9_2_8, -6.7_2_5_7], [4.4_4_2_9, 0.8_4_9_6, -5.8_5_8_5]] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" _snake_case = ViTModel.from_pretrained('''facebook/dino-vits8''' , torch_dtype=torch.floataa , device_map='''auto''' ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=__lowerCamelCase , return_tensors='''pt''' ) _snake_case = inputs.pixel_values.to(__lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): _snake_case = model(__lowerCamelCase )
103
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
0
"""simple docstring""" from collections import deque from math import floor from random import random from time import time class UpperCamelCase__ : """simple docstring""" def __init__( self ) -> Tuple: A__ = {} def snake_case__ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 ) -> Tuple: if self.graph.get(SCREAMING_SNAKE_CASE__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: A__ = [[w, v]] if not self.graph.get(SCREAMING_SNAKE_CASE__ ): A__ = [] def snake_case__ ( self ) -> Optional[Any]: return list(self.graph ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> List[Any]: if s == d: return [] A__ = [] A__ = [] if s == -2: A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(SCREAMING_SNAKE_CASE__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return visited def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-1 ) -> Union[str, Any]: if c == -1: A__ = floor(random() * 10000 ) + 10 for i in range(SCREAMING_SNAKE_CASE__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): A__ = floor(random() * c ) + 1 if n != i: self.add_pair(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 ) -> int: A__ = deque() A__ = [] if s == -2: A__ = list(self.graph )[0] d.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) while d: A__ = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def snake_case__ ( self , SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: A__ = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def snake_case__ ( self , SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: return len(self.graph[u] ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 ) -> Optional[int]: A__ = [] A__ = [] if s == -2: A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = s A__ = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return sorted_nodes def snake_case__ ( self ) -> int: A__ = [] A__ = [] A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = -2 A__ = [] A__ = s A__ = False A__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): A__ = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() A__ = True if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) A__ = s A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return list(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self ) -> int: A__ = [] A__ = [] A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = -2 A__ = [] A__ = s A__ = False A__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): A__ = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() A__ = True if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) A__ = s A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return False def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> Union[str, Any]: A__ = time() self.dfs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ = time() return end - begin def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 ) -> Dict: A__ = time() self.bfs(SCREAMING_SNAKE_CASE__ ) A__ = time() return end - begin class UpperCamelCase__ : """simple docstring""" def __init__( self ) -> Tuple: A__ = {} def snake_case__ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 ) -> Dict: # check if the u exists if self.graph.get(SCREAMING_SNAKE_CASE__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist A__ = [[w, v]] # add the other way if self.graph.get(SCREAMING_SNAKE_CASE__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist A__ = [[w, u]] def snake_case__ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Optional[int]: if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(SCREAMING_SNAKE_CASE__ ) # the other way round if self.graph.get(SCREAMING_SNAKE_CASE__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> List[str]: if s == d: return [] A__ = [] A__ = [] if s == -2: A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(SCREAMING_SNAKE_CASE__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return visited def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-1 ) -> Optional[int]: if c == -1: A__ = floor(random() * 10000 ) + 10 for i in range(SCREAMING_SNAKE_CASE__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): A__ = floor(random() * c ) + 1 if n != i: self.add_pair(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 ) -> Any: A__ = deque() A__ = [] if s == -2: A__ = list(self.graph )[0] d.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) while d: A__ = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def snake_case__ ( self , SCREAMING_SNAKE_CASE__ ) -> Dict: return len(self.graph[u] ) def snake_case__ ( self ) -> Any: A__ = [] A__ = [] A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = -2 A__ = [] A__ = s A__ = False A__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): A__ = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() A__ = True if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) A__ = s A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return list(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self ) -> Any: A__ = [] A__ = [] A__ = list(self.graph )[0] stack.append(SCREAMING_SNAKE_CASE__ ) visited.append(SCREAMING_SNAKE_CASE__ ) A__ = -2 A__ = [] A__ = s A__ = False A__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: A__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): A__ = len(SCREAMING_SNAKE_CASE__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) A__ = node[1] break # check if all the children are visited if s == ss: stack.pop() A__ = True if len(SCREAMING_SNAKE_CASE__ ) != 0: A__ = stack[len(SCREAMING_SNAKE_CASE__ ) - 1] else: A__ = False indirect_parents.append(SCREAMING_SNAKE_CASE__ ) A__ = s A__ = ss # check if se have reached the starting point if len(SCREAMING_SNAKE_CASE__ ) == 0: return False def snake_case__ ( self ) -> Any: return list(self.graph ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 , SCREAMING_SNAKE_CASE__=-1 ) -> Tuple: A__ = time() self.dfs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ = time() return end - begin def snake_case__ ( self , SCREAMING_SNAKE_CASE__=-2 ) -> str: A__ = time() self.bfs(SCREAMING_SNAKE_CASE__ ) A__ = time() return end - begin
104
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
0
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def __UpperCAmelCase ( lowerCamelCase_ : List[str] ) -> Dict: """simple docstring""" if isinstance(lowerCamelCase_ , collections.abc.Iterable ): return x return (x, x) @require_flax class lowerCAmelCase_ : def snake_case ( self ,snake_case__ ,snake_case__ ): pass def snake_case ( self ): pass def snake_case ( self ): pass def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : int = np.abs((a - b) ).max() self.assertLessEqual(snake_case__ ,snake_case__ ,F'Difference between torch and flax is {diff} (>= {tol}).' ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__=None ,**snake_case__ ): SCREAMING_SNAKE_CASE_ : Dict = VisionTextDualEncoderConfig.from_vision_text_configs(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : str = FlaxVisionTextDualEncoderModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : int = model(input_ids=snake_case__ ,pixel_values=snake_case__ ,attention_mask=snake_case__ ) self.assertEqual(output['text_embeds'].shape ,(input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['image_embeds'].shape ,(pixel_values.shape[0], config.projection_dim) ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__=None ,**snake_case__ ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : int = self.get_vision_text_model(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : List[Any] = {'vision_model': vision_model, 'text_model': text_model} SCREAMING_SNAKE_CASE_ : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**snake_case__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = model(input_ids=snake_case__ ,pixel_values=snake_case__ ,attention_mask=snake_case__ ) self.assertEqual(output['text_embeds'].shape ,(input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['image_embeds'].shape ,(pixel_values.shape[0], model.config.projection_dim) ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__=None ,**snake_case__ ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_vision_text_model(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = {'vision_model': vision_model, 'text_model': text_model} SCREAMING_SNAKE_CASE_ : str = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**snake_case__ ) SCREAMING_SNAKE_CASE_ : Any = model(input_ids=snake_case__ ,pixel_values=snake_case__ ,attention_mask=snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : str = FlaxVisionTextDualEncoderModel.from_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : int = model(input_ids=snake_case__ ,pixel_values=snake_case__ ,attention_mask=snake_case__ ) SCREAMING_SNAKE_CASE_ : int = after_output[0] SCREAMING_SNAKE_CASE_ : Tuple = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(snake_case__ ,1E-3 ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__=None ,**snake_case__ ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_vision_text_model(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : Any = {'vision_model': vision_model, 'text_model': text_model} SCREAMING_SNAKE_CASE_ : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**snake_case__ ) SCREAMING_SNAKE_CASE_ : List[str] = model( input_ids=snake_case__ ,pixel_values=snake_case__ ,attention_mask=snake_case__ ,output_attentions=snake_case__ ) SCREAMING_SNAKE_CASE_ : int = output.vision_model_output.attentions self.assertEqual(len(snake_case__ ) ,vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE_ : Tuple = to_atuple(vision_model.config.image_size ) SCREAMING_SNAKE_CASE_ : int = to_atuple(vision_model.config.patch_size ) SCREAMING_SNAKE_CASE_ : List[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) SCREAMING_SNAKE_CASE_ : str = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] ,(vision_config.num_attention_heads, seq_len, seq_len) ) SCREAMING_SNAKE_CASE_ : str = output.text_model_output.attentions self.assertEqual(len(snake_case__ ) ,text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] ,(text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) ,) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ): pt_model.to(snake_case__ ) pt_model.eval() # prepare inputs SCREAMING_SNAKE_CASE_ : int = inputs_dict SCREAMING_SNAKE_CASE_ : Tuple = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): SCREAMING_SNAKE_CASE_ : List[str] = pt_model(**snake_case__ ).to_tuple() SCREAMING_SNAKE_CASE_ : Optional[Any] = fx_model(**snake_case__ ).to_tuple() self.assertEqual(len(snake_case__ ) ,len(snake_case__ ) ,'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output in zip(fx_outputs[:4] ,pt_outputs[:4] ): self.assert_almost_equals(snake_case__ ,pt_output.numpy() ,4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : Any = FlaxVisionTextDualEncoderModel.from_pretrained(snake_case__ ,from_pt=snake_case__ ) SCREAMING_SNAKE_CASE_ : Tuple = fx_model_loaded(**snake_case__ ).to_tuple() self.assertEqual(len(snake_case__ ) ,len(snake_case__ ) ,'Output lengths differ between Flax and PyTorch' ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] ,pt_outputs[:4] ): self.assert_almost_equals(snake_case__ ,pt_output.numpy() ,4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : str = VisionTextDualEncoderModel.from_pretrained(snake_case__ ,from_flax=snake_case__ ) pt_model_loaded.to(snake_case__ ) pt_model_loaded.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ : Union[str, Any] = pt_model_loaded(**snake_case__ ).to_tuple() self.assertEqual(len(snake_case__ ) ,len(snake_case__ ) ,'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] ,pt_outputs_loaded[:4] ): self.assert_almost_equals(snake_case__ ,pt_output_loaded.numpy() ,4E-2 ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : List[str] = VisionTextDualEncoderConfig.from_vision_text_configs(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : int = VisionTextDualEncoderModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = FlaxVisionTextDualEncoderModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() ,snake_case__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = fx_state self.check_pt_flax_equivalence(snake_case__ ,snake_case__ ,snake_case__ ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : Dict = VisionTextDualEncoderConfig.from_vision_text_configs(snake_case__ ,snake_case__ ) SCREAMING_SNAKE_CASE_ : List[Any] = VisionTextDualEncoderModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = FlaxVisionTextDualEncoderModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : Tuple = load_flax_weights_in_pytorch_model(snake_case__ ,fx_model.params ) self.check_pt_flax_equivalence(snake_case__ ,snake_case__ ,snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Optional[Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Tuple = self.prepare_config_and_inputs() self.check_save_load(**snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : int = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**snake_case__ ) @is_pt_flax_cross_test def snake_case ( self ): SCREAMING_SNAKE_CASE_ : str = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ : Union[str, Any] = config_inputs_dict.pop('vision_config' ) SCREAMING_SNAKE_CASE_ : Tuple = config_inputs_dict.pop('text_config' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = config_inputs_dict self.check_equivalence_pt_to_flax(snake_case__ ,snake_case__ ,snake_case__ ) self.check_equivalence_flax_to_pt(snake_case__ ,snake_case__ ,snake_case__ ) @slow def snake_case ( self ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_pretrained_model_and_inputs() SCREAMING_SNAKE_CASE_ : Any = model_a(**snake_case__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(snake_case__ ) SCREAMING_SNAKE_CASE_ : List[str] = model_a(**snake_case__ ) SCREAMING_SNAKE_CASE_ : int = after_outputs[0] SCREAMING_SNAKE_CASE_ : Union[str, Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(snake_case__ ,1E-5 ) @require_flax class lowerCAmelCase_ ( lowerCamelCase_ , unittest.TestCase ): def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-vit' ,'hf-internal-testing/tiny-bert' ,vision_from_pt=snake_case__ ,text_from_pt=snake_case__ ,) SCREAMING_SNAKE_CASE_ : List[Any] = 13 SCREAMING_SNAKE_CASE_ : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) SCREAMING_SNAKE_CASE_ : Any = ids_tensor([batch_size, 4] ,model.config.text_config.vocab_size ) SCREAMING_SNAKE_CASE_ : Optional[int] = random_attention_mask([batch_size, 4] ) SCREAMING_SNAKE_CASE_ : Tuple = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def snake_case ( self ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : int = FlaxViTModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : str = FlaxBertModel(snake_case__ ) return vision_model, text_model def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Dict = FlaxViTModelTester(self ) SCREAMING_SNAKE_CASE_ : int = FlaxBertModelTester(self ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = vit_model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ : List[str] = bert_model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Tuple = vision_config_and_inputs SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class lowerCAmelCase_ ( lowerCamelCase_ , unittest.TestCase ): def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Tuple = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-clip' ,'hf-internal-testing/tiny-bert' ,vision_from_pt=snake_case__ ,text_from_pt=snake_case__ ,) SCREAMING_SNAKE_CASE_ : Any = 13 SCREAMING_SNAKE_CASE_ : str = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) SCREAMING_SNAKE_CASE_ : Dict = ids_tensor([batch_size, 4] ,model.config.text_config.vocab_size ) SCREAMING_SNAKE_CASE_ : int = random_attention_mask([batch_size, 4] ) SCREAMING_SNAKE_CASE_ : Any = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def snake_case ( self ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : int = FlaxCLIPVisionModel(snake_case__ ) SCREAMING_SNAKE_CASE_ : Dict = FlaxBertModel(snake_case__ ) return vision_model, text_model def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Tuple = FlaxCLIPVisionModelTester(self ) SCREAMING_SNAKE_CASE_ : Optional[Any] = FlaxBertModelTester(self ) SCREAMING_SNAKE_CASE_ : Optional[int] = clip_model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ : Tuple = bert_model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[Any] = vision_config_and_inputs SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def snake_case ( self ): SCREAMING_SNAKE_CASE_ : int = FlaxVisionTextDualEncoderModel.from_pretrained('clip-italian/clip-italian' ,logit_scale_init_value=1.0 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = VisionTextDualEncoderProcessor.from_pretrained('clip-italian/clip-italian' ) SCREAMING_SNAKE_CASE_ : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) SCREAMING_SNAKE_CASE_ : Optional[int] = processor( text=['una foto di un gatto', 'una foto di un cane'] ,images=snake_case__ ,padding=snake_case__ ,return_tensors='np' ) SCREAMING_SNAKE_CASE_ : List[Any] = model(**snake_case__ ) # verify the logits self.assertEqual(outputs.logits_per_image.shape ,(inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape ,(inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) ,) SCREAMING_SNAKE_CASE_ : Dict = np.array([[1.2284727, 0.3104122]] ) self.assertTrue(np.allclose(outputs.logits_per_image ,snake_case__ ,atol=1E-3 ) )
105
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
0
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class lowerCAmelCase__ ( nn.Module ): def __init__( self : List[Any] ) -> Tuple: super().__init__() A = nn.Linear(3 , 4 ) A = nn.BatchNormad(4 ) A = nn.Linear(4 , 5 ) def __UpperCamelCase ( self : Union[str, Any] , __UpperCamelCase : List[Any] ) -> str: return self.lineara(self.batchnorm(self.lineara(__UpperCamelCase ) ) ) class lowerCAmelCase__ ( _lowerCamelCase ): def __UpperCamelCase ( self : Any , __UpperCamelCase : Any , *__UpperCamelCase : Any , **__UpperCamelCase : List[str] ) -> Dict: return (args[0] + 1,) + args[1:], kwargs class lowerCAmelCase__ ( _lowerCamelCase ): def __UpperCamelCase ( self : Tuple , __UpperCamelCase : Dict , __UpperCamelCase : Tuple ) -> str: return output + 1 class lowerCAmelCase__ ( unittest.TestCase ): def __UpperCamelCase ( self : Dict ) -> Any: A = ModelForTest() A = ModelHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) self.assertEqual(test_model._hf_hook , __UpperCamelCase ) self.assertTrue(hasattr(__UpperCamelCase , '_old_forward' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , 'forward' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['x'] ) remove_hook_from_module(__UpperCamelCase ) self.assertFalse(hasattr(__UpperCamelCase , '_hf_hook' ) ) self.assertFalse(hasattr(__UpperCamelCase , '_old_forward' ) ) def __UpperCamelCase ( self : Any ) -> Union[str, Any]: A = ModelForTest() A = ModelHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) add_hook_to_module(__UpperCamelCase , __UpperCamelCase , append=__UpperCamelCase ) self.assertEqual(isinstance(test_model._hf_hook , __UpperCamelCase ) , __UpperCamelCase ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(__UpperCamelCase , '_old_forward' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , 'forward' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['x'] ) remove_hook_from_module(__UpperCamelCase ) self.assertFalse(hasattr(__UpperCamelCase , '_hf_hook' ) ) self.assertFalse(hasattr(__UpperCamelCase , '_old_forward' ) ) def __UpperCamelCase ( self : List[Any] ) -> Optional[Any]: A = ModelForTest() A = torch.randn(2 , 3 ) A = test_model(x + 1 ) A = test_model(x + 2 ) A = PreForwardHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) self.assertTrue(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain A = PreForwardHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) self.assertTrue(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks A = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) assert torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1e-5 ) def __UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: A = ModelForTest() A = torch.randn(2 , 3 ) A = test_model(__UpperCamelCase ) A = PostForwardHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) self.assertTrue(torch.allclose(__UpperCamelCase , output + 1 , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain A = PostForwardHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) self.assertTrue(torch.allclose(__UpperCamelCase , output + 1 , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks A = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) assert torch.allclose(__UpperCamelCase , output + 2 , atol=1e-5 ) def __UpperCamelCase ( self : Optional[int] ) -> List[str]: A = ModelForTest() A = torch.randn(2 , 3 ) A = test_model(__UpperCamelCase ) A = PostForwardHook() add_hook_to_module(__UpperCamelCase , __UpperCamelCase ) A = test_model(__UpperCamelCase ) self.assertTrue(torch.allclose(__UpperCamelCase , output + 1 ) ) self.assertTrue(outputa.requires_grad ) A = True A = test_model(__UpperCamelCase ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def __UpperCamelCase ( self : List[Any] ) -> List[Any]: A = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(__UpperCamelCase , AlignDevicesHook(io_same_device=__UpperCamelCase ) ) A = torch.randn(2 , 3 ).to(0 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , torch.device(0 ) ) def __UpperCamelCase ( self : Optional[int] ) -> Optional[int]: A = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices A = {'execution_device': 0 if torch.cuda.is_available() else 'cpu', 'offload': True} add_hook_to_module(model.lineara , AlignDevicesHook(**__UpperCamelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**__UpperCamelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**__UpperCamelCase ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device A = torch.device(hook_kwargs['execution_device'] ) self.assertEqual(model.batchnorm.running_mean.device , __UpperCamelCase ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload A = { 'execution_device': 0 if torch.cuda.is_available() else 'cpu', 'offload': True, 'offload_buffers': True, } add_hook_to_module(model.lineara , AlignDevicesHook(**__UpperCamelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**__UpperCamelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**__UpperCamelCase ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) def __UpperCamelCase ( self : Any ) -> Optional[Any]: A = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices A = 0 if torch.cuda.is_available() else 'cpu' attach_align_device_hook(__UpperCamelCase , execution_device=__UpperCamelCase , offload=__UpperCamelCase ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device A = torch.device(__UpperCamelCase ) self.assertEqual(model.batchnorm.running_mean.device , __UpperCamelCase ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(__UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload attach_align_device_hook(__UpperCamelCase , execution_device=__UpperCamelCase , offload=__UpperCamelCase , offload_buffers=__UpperCamelCase ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(__UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) def __UpperCamelCase ( self : str ) -> List[Any]: A = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices A = 0 if torch.cuda.is_available() else 'cpu' attach_align_device_hook( __UpperCamelCase , execution_device=__UpperCamelCase , offload=__UpperCamelCase , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device A = torch.device(__UpperCamelCase ) self.assertEqual(model.batchnorm.running_mean.device , __UpperCamelCase ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(__UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload attach_align_device_hook( __UpperCamelCase , execution_device=__UpperCamelCase , offload=__UpperCamelCase , weights_map=model.state_dict() , offload_buffers=__UpperCamelCase , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) A = torch.randn(2 , 3 ) A = model(__UpperCamelCase ) self.assertEqual(output.device , __UpperCamelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(__UpperCamelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) )
106
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class lowercase_ ( _UpperCamelCase ): """simple docstring""" __lowerCAmelCase = "facebook/bart-large-mnli" __lowerCAmelCase = ( "This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which " "should be the text to classify, and `labels`, which should be the list of labels to use for classification. " "It returns the most likely label in the list of provided `labels` for the input text." ) __lowerCAmelCase = "text_classifier" __lowerCAmelCase = AutoTokenizer __lowerCAmelCase = AutoModelForSequenceClassification __lowerCAmelCase = ["text", ["text"]] __lowerCAmelCase = ["text"] def __UpperCAmelCase ( self : List[str] ) -> str: super().setup() _A = self.model.config _A = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('entail' ): _A = int(UpperCamelCase__ ) if self.entailment_id == -1: raise ValueError('Could not determine the entailment ID from the model config, please pass it at init.' ) def __UpperCAmelCase ( self : List[str], UpperCamelCase__ : Optional[Any], UpperCamelCase__ : int ) -> Optional[Any]: _A = labels return self.pre_processor( [text] * len(UpperCamelCase__ ), [f'This example is {label}' for label in labels], return_tensors='pt', padding='max_length', ) def __UpperCAmelCase ( self : int, UpperCamelCase__ : int ) -> Dict: _A = outputs.logits _A = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
107
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): '''simple docstring''' _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def lowerCamelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" torch.manual_seed(0 ) _UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase , ) _UpperCAmelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase ) torch.manual_seed(0 ) _UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="""gelu""" , projection_dim=512 , ) _UpperCAmelCase = CLIPTextModel(lowerCamelCase ) _UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCamelCase ( self : Optional[int] , lowerCamelCase : Optional[int] , lowerCamelCase : Optional[Any]=0 ) -> Optional[int]: """simple docstring""" # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched _UpperCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) _UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _UpperCAmelCase = Image.fromarray(np.uinta(lowerCamelCase ) ).convert("""RGB""" ).resize((64, 64) ) _UpperCAmelCase = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((64, 64) ) if str(lowerCamelCase ).startswith("""mps""" ): _UpperCAmelCase = torch.manual_seed(lowerCamelCase ) else: _UpperCAmelCase = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) _UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCamelCase ( self : Optional[int] ) -> Any: """simple docstring""" _UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = StableDiffusionInpaintPipeline(**lowerCamelCase ) _UpperCAmelCase = sd_pipe.to(lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase ) _UpperCAmelCase = self.get_dummy_inputs(lowerCamelCase ) _UpperCAmelCase = sd_pipe(**lowerCamelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self : Tuple ) -> Tuple: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase ( self : Dict ) -> int: """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : Any ) -> List[Any]: """simple docstring""" _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) _UpperCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) _UpperCAmelCase = """stabilityai/stable-diffusion-2-inpainting""" _UpperCAmelCase = StableDiffusionInpaintPipeline.from_pretrained(lowerCamelCase , safety_checker=lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() _UpperCAmelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , output_type="""np""" , ) _UpperCAmelCase = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def lowerCamelCase ( self : int ) -> str: """simple docstring""" _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) _UpperCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) _UpperCAmelCase = """stabilityai/stable-diffusion-2-inpainting""" _UpperCAmelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase , torch_dtype=torch.floataa , safety_checker=lowerCamelCase , ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() _UpperCAmelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , output_type="""np""" , ) _UpperCAmelCase = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def lowerCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) _UpperCAmelCase = """stabilityai/stable-diffusion-2-inpainting""" _UpperCAmelCase = PNDMScheduler.from_pretrained(lowerCamelCase , subfolder="""scheduler""" ) _UpperCAmelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase , safety_checker=lowerCamelCase , scheduler=lowerCamelCase , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _UpperCAmelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , num_inference_steps=2 , output_type="""np""" , ) _UpperCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
108
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
0
'''simple docstring''' from PIL import Image def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Image: '''simple docstring''' __SCREAMING_SNAKE_CASE = (259 * (level + 255)) / (255 * (259 - level)) def contrast(__UpperCAmelCase ) -> int: return int(128 + factor * (c - 128) ) return img.point(__UpperCAmelCase ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change contrast to 170 a = change_contrast(img, 170) cont_img.save("image_data/lena_high_contrast.png", format="png")
109
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
0
"""simple docstring""" from statistics import mean, stdev def lowerCamelCase ( _snake_case ,_snake_case = 3 ): UpperCAmelCase__ : Tuple = min(_snake_case ) UpperCAmelCase__ : Any = max(_snake_case ) # normalize data return [round((x - x_min) / (x_max - x_min) ,_snake_case ) for x in data] def lowerCamelCase ( _snake_case ,_snake_case = 3 ): UpperCAmelCase__ : Optional[Any] = mean(_snake_case ) UpperCAmelCase__ : Optional[int] = stdev(_snake_case ) # standardize data return [round((x - mu) / (sigma) ,_snake_case ) for x in data]
110
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('1.0.0a'): raise Exception('requires fairseq >= 1.0.0a') logging.set_verbosity_info() __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """Hello world! cécé herlolip""" def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = FairseqRobertaModel.from_pretrained(_SCREAMING_SNAKE_CASE ) roberta.eval() # disable dropout _snake_case = roberta.model.encoder.sentence_encoder _snake_case = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1E-5 , ) if classification_head: _snake_case = roberta.model.classification_heads["""mnli"""].out_proj.weight.shape[0] print("""Our RoBERTa config:""" , _SCREAMING_SNAKE_CASE ) _snake_case = XLMRobertaXLForSequenceClassification(_SCREAMING_SNAKE_CASE ) if classification_head else XLMRobertaXLForMaskedLM(_SCREAMING_SNAKE_CASE ) model.eval() # Now let's copy all the weights. # Embeddings _snake_case = roberta_sent_encoder.embed_tokens.weight _snake_case = roberta_sent_encoder.embed_positions.weight _snake_case = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. _snake_case = roberta_sent_encoder.layer_norm.weight _snake_case = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer _snake_case = model.roberta.encoder.layer[i] _snake_case = roberta_sent_encoder.layers[i] _snake_case = layer.attention _snake_case = roberta_layer.self_attn_layer_norm.weight _snake_case = roberta_layer.self_attn_layer_norm.bias # self attention _snake_case = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) _snake_case = roberta_layer.self_attn.q_proj.weight _snake_case = roberta_layer.self_attn.q_proj.bias _snake_case = roberta_layer.self_attn.k_proj.weight _snake_case = roberta_layer.self_attn.k_proj.bias _snake_case = roberta_layer.self_attn.v_proj.weight _snake_case = roberta_layer.self_attn.v_proj.bias # self-attention output _snake_case = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape _snake_case = roberta_layer.self_attn.out_proj.weight _snake_case = roberta_layer.self_attn.out_proj.bias # this one is final layer norm _snake_case = roberta_layer.final_layer_norm.weight _snake_case = roberta_layer.final_layer_norm.bias # intermediate _snake_case = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape _snake_case = roberta_layer.fca.weight _snake_case = roberta_layer.fca.bias # output _snake_case = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape _snake_case = roberta_layer.fca.weight _snake_case = roberta_layer.fca.bias # end of layer if classification_head: _snake_case = roberta.model.classification_heads["""mnli"""].dense.weight _snake_case = roberta.model.classification_heads["""mnli"""].dense.bias _snake_case = roberta.model.classification_heads["""mnli"""].out_proj.weight _snake_case = roberta.model.classification_heads["""mnli"""].out_proj.bias else: # LM Head _snake_case = roberta.model.encoder.lm_head.dense.weight _snake_case = roberta.model.encoder.lm_head.dense.bias _snake_case = roberta.model.encoder.lm_head.layer_norm.weight _snake_case = roberta.model.encoder.lm_head.layer_norm.bias _snake_case = roberta.model.encoder.lm_head.weight _snake_case = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. _snake_case = roberta.encode(_SCREAMING_SNAKE_CASE ).unsqueeze(0 ) # batch of size 1 _snake_case = model(_SCREAMING_SNAKE_CASE )[0] if classification_head: _snake_case = roberta.model.classification_heads["""mnli"""](roberta.extract_features(_SCREAMING_SNAKE_CASE ) ) else: _snake_case = roberta.model(_SCREAMING_SNAKE_CASE )[0] print(our_output.shape , their_output.shape ) _snake_case = torch.max(torch.abs(our_output - their_output ) ).item() print(f"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 _snake_case = torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) print("""Do both models output the same tensors?""" , """🔥""" if success else """💩""" ) if not success: raise Exception("""Something went wRoNg""" ) pathlib.Path(_SCREAMING_SNAKE_CASE ).mkdir(parents=_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--roberta_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) __lowerCAmelCase = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
585
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
0
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase_ = False class a ( unittest.TestCase ): pass @nightly @require_torch_gpu class a ( unittest.TestCase ): def UpperCAmelCase__ ( self : Any ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : Dict ): """simple docstring""" __lowerCAmelCase = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pipe.dual_guided( prompt="first prompt" , image=_lowerCAmelCase , text_to_image_strength=0.7_5 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowerCAmelCase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = generator.manual_seed(0 ) __lowerCAmelCase = pipe.dual_guided( prompt="first prompt" , image=_lowerCAmelCase , text_to_image_strength=0.7_5 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def UpperCAmelCase__ ( self : Any ): """simple docstring""" __lowerCAmelCase = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = "cyberpunk 2077" __lowerCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.7_5 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images __lowerCAmelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCAmelCase = "A painting of a squirrel eating a burger " __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images __lowerCAmelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCAmelCase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="numpy" ).images __lowerCAmelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
611
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
0
'''simple docstring''' import sys from collections import defaultdict class _a : def __init__( self ) -> Tuple: _snake_case = [] def _lowercase ( self ,_SCREAMING_SNAKE_CASE ) -> Dict: return self.node_position[vertex] def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> List[str]: _snake_case = pos def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _snake_case = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _snake_case = 2 * start + 1 else: _snake_case = 2 * start + 2 if heap[smallest_child] < heap[start]: _snake_case , _snake_case = heap[smallest_child], positions[smallest_child] _snake_case , _snake_case = ( heap[start], positions[start], ) _snake_case , _snake_case = temp, tempa _snake_case = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] ,self.get_position(positions[start] ) ) self.set_position(positions[start] ,_lowerCAmelCase ) self.top_to_bottom(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: _snake_case = position[index] while index != 0: _snake_case = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: _snake_case = heap[parent] _snake_case = position[parent] self.set_position(position[parent] ,_lowerCAmelCase ) else: _snake_case = val _snake_case = temp self.set_position(_lowerCAmelCase ,_lowerCAmelCase ) break _snake_case = parent else: _snake_case = val _snake_case = temp self.set_position(_lowerCAmelCase ,0 ) def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: _snake_case = len(_lowerCAmelCase ) // 2 - 1 for i in range(_lowerCAmelCase ,-1 ,-1 ): self.top_to_bottom(_lowerCAmelCase ,_lowerCAmelCase ,len(_lowerCAmelCase ) ,_lowerCAmelCase ) def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> str: _snake_case = positions[0] _snake_case = sys.maxsize self.top_to_bottom(_lowerCAmelCase ,0 ,len(_lowerCAmelCase ) ,_lowerCAmelCase ) return temp def __a ( _UpperCamelCase: List[Any] ) -> int: """simple docstring""" _snake_case = Heap() _snake_case = [0] * len(_UpperCamelCase ) _snake_case = [-1] * len(_UpperCamelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _snake_case = [] # Heap of Distance of vertices from their neighboring vertex _snake_case = [] for vertex in range(len(_UpperCamelCase ) ): distance_tv.append(sys.maxsize ) positions.append(_UpperCamelCase ) heap.node_position.append(_UpperCamelCase ) _snake_case = [] _snake_case = 1 _snake_case = sys.maxsize for neighbor, distance in adjacency_list[0]: _snake_case = 0 _snake_case = distance heap.heapify(_UpperCamelCase , _UpperCamelCase ) for _ in range(1 , len(_UpperCamelCase ) ): _snake_case = heap.delete_minimum(_UpperCamelCase , _UpperCamelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _snake_case = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(_UpperCamelCase )] ): _snake_case = distance heap.bottom_to_top( _UpperCamelCase , heap.get_position(_UpperCamelCase ) , _UpperCamelCase , _UpperCamelCase ) _snake_case = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > UpperCamelCase_ : int = int(input('''Enter number of edges: ''').strip()) UpperCamelCase_ : Optional[Any] = defaultdict(list) for _ in range(edges_number): UpperCamelCase_ : List[Any] = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
185
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
0
"""simple docstring""" import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def SCREAMING_SNAKE_CASE_ ( snake_case : Tuple = 8 )-> str: _lowerCamelCase = ascii_letters + digits + punctuation return "".join(secrets.choice(snake_case ) for _ in range(snake_case ) ) def SCREAMING_SNAKE_CASE_ ( snake_case : Any , snake_case : Dict )-> Union[str, Any]: i -= len(snake_case ) _lowerCamelCase = i // 3 _lowerCamelCase = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) _lowerCamelCase = ( chars_incl + random(snake_case , quotient + remainder ) + random(snake_case , snake_case ) + random(snake_case , snake_case ) ) _lowerCamelCase = list(snake_case ) shuffle(snake_case ) return "".join(snake_case ) # random is a generalised function for letters, characters and numbers def SCREAMING_SNAKE_CASE_ ( snake_case : List[Any] , snake_case : List[str] )-> Union[str, Any]: return "".join(secrets.choice(snake_case ) for _ in range(snake_case ) ) def SCREAMING_SNAKE_CASE_ ( snake_case : List[Any] , snake_case : Dict )-> Optional[Any]: pass # Put your code here... def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[int] , snake_case : List[str] )-> int: pass # Put your code here... def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[Any] , snake_case : Tuple )-> int: pass # Put your code here... def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[int] , snake_case : Optional[Any] = 8 )-> Optional[int]: if len(snake_case ) < min_length: # Your Password must be at least 8 characters long return False _lowerCamelCase = any(char in ascii_uppercase for char in password ) _lowerCamelCase = any(char in ascii_lowercase for char in password ) _lowerCamelCase = any(char in digits for char in password ) _lowerCamelCase = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def SCREAMING_SNAKE_CASE_ ( )-> str: _lowerCamelCase = int(input('Please indicate the max length of your password: ' ).strip() ) _lowerCamelCase = input( 'Please indicate the characters that must be in your password: ' ).strip() print('Password generated:' , password_generator(snake_case ) ) print( 'Alternative Password generated:' , alternative_password_generator(snake_case , snake_case ) , ) print('[If you are thinking of using this passsword, You better save it.]' ) if __name__ == "__main__": main()
650
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class A_ ( unittest.TestCase ): _UpperCAmelCase : Union[str, Any] = StableDiffusionLDMaDPipeline _UpperCAmelCase : List[str] = TEXT_TO_IMAGE_PARAMS _UpperCAmelCase : List[str] = TEXT_TO_IMAGE_BATCH_PARAMS _UpperCAmelCase : str = TEXT_TO_IMAGE_IMAGE_PARAMS def lowerCAmelCase ( self : Tuple): torch.manual_seed(0) __lowerCamelCase : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) ,layers_per_block=2 ,sample_size=3_2 ,in_channels=4 ,out_channels=4 ,down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') ,up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') ,cross_attention_dim=3_2 ,) __lowerCamelCase : Tuple = DDIMScheduler( beta_start=0.00085 ,beta_end=0.012 ,beta_schedule='scaled_linear' ,clip_sample=_lowerCAmelCase ,set_alpha_to_one=_lowerCAmelCase ,) torch.manual_seed(0) __lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] ,in_channels=6 ,out_channels=6 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=4 ,) torch.manual_seed(0) __lowerCamelCase : str = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=3_2 ,intermediate_size=3_7 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_0_0_0 ,) __lowerCamelCase : int = CLIPTextModel(_lowerCAmelCase) __lowerCamelCase : List[str] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') __lowerCamelCase : Optional[Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowerCAmelCase ( self : str ,SCREAMING_SNAKE_CASE__ : Any ,SCREAMING_SNAKE_CASE__ : Dict=0): if str(_lowerCAmelCase).startswith('mps'): __lowerCamelCase : Optional[int] = torch.manual_seed(_lowerCAmelCase) else: __lowerCamelCase : Any = torch.Generator(device=_lowerCAmelCase).manual_seed(_lowerCAmelCase) __lowerCamelCase : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def lowerCAmelCase ( self : List[Any]): __lowerCamelCase : str = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : List[str] = self.get_dummy_components() __lowerCamelCase : List[str] = StableDiffusionLDMaDPipeline(**_lowerCAmelCase) __lowerCamelCase : Tuple = ldmad_pipe.to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : List[str] = self.get_dummy_inputs(_lowerCAmelCase) __lowerCamelCase : Optional[int] = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Tuple = output.rgb, output.depth __lowerCamelCase : Any = rgb[0, -3:, -3:, -1] __lowerCamelCase : Optional[int] = depth[0, -3:, -1] assert rgb.shape == (1, 6_4, 6_4, 3) assert depth.shape == (1, 6_4, 6_4) __lowerCamelCase : int = np.array( [0.37338176, 0.70247, 0.74203193, 0.51643604, 0.58256793, 0.60932136, 0.4181095, 0.48355877, 0.46535262]) __lowerCamelCase : Optional[int] = np.array([103.46727, 85.812004, 87.849236]) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb).max() < 1E-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth).max() < 1E-2 def lowerCAmelCase ( self : Tuple): __lowerCamelCase : Optional[Any] = self.get_dummy_components() __lowerCamelCase : Optional[int] = StableDiffusionLDMaDPipeline(**_lowerCAmelCase) __lowerCamelCase : Any = ldmad_pipe.to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Optional[int] = self.get_dummy_inputs(_lowerCAmelCase) __lowerCamelCase : Union[str, Any] = 3 * [inputs['prompt']] # forward __lowerCamelCase : List[Any] = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : str = output.rgb, output.depth __lowerCamelCase : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] __lowerCamelCase : Optional[Any] = depth_slice_a[0, -3:, -1] __lowerCamelCase : List[str] = self.get_dummy_inputs(_lowerCAmelCase) __lowerCamelCase : Any = 3 * [inputs.pop('prompt')] __lowerCamelCase : Dict = ldmad_pipe.tokenizer( _lowerCAmelCase ,padding='max_length' ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=_lowerCAmelCase ,return_tensors='pt' ,) __lowerCamelCase : str = text_inputs['input_ids'].to(_lowerCAmelCase) __lowerCamelCase : Dict = ldmad_pipe.text_encoder(_lowerCAmelCase)[0] __lowerCamelCase : Tuple = prompt_embeds # forward __lowerCamelCase : Dict = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Any = output.rgb, output.depth __lowerCamelCase : str = rgb_slice_a[0, -3:, -3:, -1] __lowerCamelCase : Optional[int] = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten()).max() < 1E-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten()).max() < 1E-4 def lowerCAmelCase ( self : Any): __lowerCamelCase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Optional[int] = self.get_dummy_components() __lowerCamelCase : Optional[int] = PNDMScheduler(skip_prk_steps=_lowerCAmelCase) __lowerCamelCase : Union[str, Any] = StableDiffusionLDMaDPipeline(**_lowerCAmelCase) __lowerCamelCase : Optional[Any] = ldmad_pipe.to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Union[str, Any] = self.get_dummy_inputs(_lowerCAmelCase) __lowerCamelCase : List[str] = 'french fries' __lowerCamelCase : int = ldmad_pipe(**_lowerCAmelCase ,negative_prompt=_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Union[str, Any] = output.rgb, output.depth __lowerCamelCase : Tuple = rgb[0, -3:, -3:, -1] __lowerCamelCase : int = depth[0, -3:, -1] assert rgb.shape == (1, 6_4, 6_4, 3) assert depth.shape == (1, 6_4, 6_4) __lowerCamelCase : Optional[Any] = np.array( [0.37044, 0.71811503, 0.7223251, 0.48603675, 0.5638391, 0.6364948, 0.42833704, 0.4901315, 0.47926217]) __lowerCamelCase : Optional[Any] = np.array([107.84738, 84.62802, 89.962135]) assert np.abs(rgb_slice.flatten() - expected_slice_rgb).max() < 1E-2 assert np.abs(depth_slice.flatten() - expected_slice_depth).max() < 1E-2 @slow @require_torch_gpu class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : int): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase ( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : int ,SCREAMING_SNAKE_CASE__ : List[Any]="cpu" ,SCREAMING_SNAKE_CASE__ : int=torch.floataa ,SCREAMING_SNAKE_CASE__ : int=0): __lowerCamelCase : int = torch.Generator(device=_lowerCAmelCase).manual_seed(_lowerCAmelCase) __lowerCamelCase : List[str] = np.random.RandomState(_lowerCAmelCase).standard_normal((1, 4, 6_4, 6_4)) __lowerCamelCase : List[str] = torch.from_numpy(_lowerCAmelCase).to(device=_lowerCAmelCase ,dtype=_lowerCAmelCase) __lowerCamelCase : int = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowerCAmelCase ( self : List[Any]): __lowerCamelCase : List[str] = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d') __lowerCamelCase : Optional[int] = ldmad_pipe.to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Optional[int] = self.get_inputs(_lowerCAmelCase) __lowerCamelCase : Dict = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Any = output.rgb, output.depth __lowerCamelCase : Tuple = rgb[0, -3:, -3:, -1].flatten() __lowerCamelCase : Tuple = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_1_2, 5_1_2, 3) assert depth.shape == (1, 5_1_2, 5_1_2) __lowerCamelCase : Any = np.array( [0.53805465, 0.56707305, 0.5486515, 0.57012236, 0.5814511, 0.56253487, 0.54843014, 0.55092263, 0.6459706]) __lowerCamelCase : List[Any] = np.array( [0.9263781, 0.6678672, 0.5486515, 0.92202145, 0.67831135, 0.56253487, 0.9241694, 0.7551478, 0.6459706]) assert np.abs(rgb_slice - expected_slice_rgb).max() < 3E-3 assert np.abs(depth_slice - expected_slice_depth).max() < 3E-3 @nightly @require_torch_gpu class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : int): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase ( self : Any ,SCREAMING_SNAKE_CASE__ : Optional[Any] ,SCREAMING_SNAKE_CASE__ : Optional[Any]="cpu" ,SCREAMING_SNAKE_CASE__ : Dict=torch.floataa ,SCREAMING_SNAKE_CASE__ : Optional[int]=0): __lowerCamelCase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase).manual_seed(_lowerCAmelCase) __lowerCamelCase : int = np.random.RandomState(_lowerCAmelCase).standard_normal((1, 4, 6_4, 6_4)) __lowerCamelCase : Any = torch.from_numpy(_lowerCAmelCase).to(device=_lowerCAmelCase ,dtype=_lowerCAmelCase) __lowerCamelCase : str = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 5_0, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowerCAmelCase ( self : Optional[Any]): __lowerCamelCase : Dict = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d').to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Any = self.get_inputs(_lowerCAmelCase) __lowerCamelCase : Optional[int] = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Optional[int] = output.rgb, output.depth __lowerCamelCase : List[Any] = 0.495586 __lowerCamelCase : str = 0.33795515 __lowerCamelCase : Tuple = 112.48518 __lowerCamelCase : Tuple = 98.489746 assert np.abs(expected_rgb_mean - rgb.mean()) < 1E-3 assert np.abs(expected_rgb_std - rgb.std()) < 1E-3 assert np.abs(expected_depth_mean - depth.mean()) < 1E-3 assert np.abs(expected_depth_std - depth.std()) < 1E-3 def lowerCAmelCase ( self : Tuple): __lowerCamelCase : Union[str, Any] = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d-4c').to(_lowerCAmelCase) ldmad_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Any = self.get_inputs(_lowerCAmelCase) __lowerCamelCase : Union[str, Any] = ldmad_pipe(**_lowerCAmelCase) __lowerCamelCase , __lowerCamelCase : Dict = output.rgb, output.depth __lowerCamelCase : Union[str, Any] = 0.4194127 __lowerCamelCase : Union[str, Any] = 0.35375586 __lowerCamelCase : List[str] = 0.5638502 __lowerCamelCase : str = 0.34686103 assert rgb.shape == (1, 5_1_2, 5_1_2, 3) assert depth.shape == (1, 5_1_2, 5_1_2, 1) assert np.abs(expected_rgb_mean - rgb.mean()) < 1E-3 assert np.abs(expected_rgb_std - rgb.std()) < 1E-3 assert np.abs(expected_depth_mean - depth.mean()) < 1E-3 assert np.abs(expected_depth_std - depth.std()) < 1E-3
652
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
0
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def snake_case ( lowerCAmelCase_=None ) -> List[Any]: if subparsers is not None: _snake_case = subparsers.add_parser('''test''' ) else: _snake_case = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=lowerCAmelCase_ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase_ ) return parser def snake_case ( lowerCAmelCase_ ) -> List[Any]: _snake_case = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: _snake_case = script_name else: _snake_case = f"""--config_file={args.config_file} {script_name}""" _snake_case = ['''accelerate-launch'''] + test_args.split() _snake_case = execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def snake_case ( ) -> Tuple: _snake_case = test_command_parser() _snake_case = parser.parse_args() test_command(lowerCAmelCase_ ) if __name__ == "__main__": main()
103
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase_ = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
458
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[Any]: if not nums: raise ValueError('List is empty' ) return sum(_UpperCAmelCase ) / len(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
295
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
0
def __UpperCamelCase ( _lowerCAmelCase = 1000 ) -> Tuple: """simple docstring""" A : str = 2**power A : Tuple = str(_lowerCAmelCase ) A : str = list(_lowerCAmelCase ) A : Optional[int] = 0 for i in list_num: sum_of_num += int(_lowerCAmelCase ) return sum_of_num if __name__ == "__main__": SCREAMING_SNAKE_CASE_:int = int(input("""Enter the power of 2: """).strip()) print("""2 ^ """, power, """ = """, 2**power) SCREAMING_SNAKE_CASE_:int = solution(power) print("""Sum of the digits is: """, result)
662
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
0
import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def snake_case_ ( _SCREAMING_SNAKE_CASE ): __lowercase = VideoMAEConfig() set_architecture_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if "finetuned" not in model_name: __lowercase = False if "finetuned" in model_name: __lowercase = "huggingface/label-files" if "kinetics" in model_name: __lowercase = 4_0_0 __lowercase = "kinetics400-id2label.json" elif "ssv2" in model_name: __lowercase = 1_7_4 __lowercase = "something-something-v2-id2label.json" else: raise ValueError("Model name should either contain 'kinetics' or 'ssv2' in case it's fine-tuned." ) __lowercase = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type="dataset" ) , "r" ) ) __lowercase = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} return config def snake_case_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if "small" in model_name: __lowercase = 3_8_4 __lowercase = 1_5_3_6 __lowercase = 1_2 __lowercase = 1_6 __lowercase = 1_2 __lowercase = 3 __lowercase = 1_9_2 __lowercase = 7_6_8 elif "large" in model_name: __lowercase = 1_0_2_4 __lowercase = 4_0_9_6 __lowercase = 2_4 __lowercase = 1_6 __lowercase = 1_2 __lowercase = 8 __lowercase = 5_1_2 __lowercase = 2_0_4_8 elif "huge" in model_name: __lowercase = 1_2_8_0 __lowercase = 5_1_2_0 __lowercase = 3_2 __lowercase = 1_6 __lowercase = 1_2 __lowercase = 8 __lowercase = 6_4_0 __lowercase = 2_5_6_0 elif "base" not in model_name: raise ValueError("Model name should include either \"small\", \"base\", \"large\", or \"huge\"" ) def snake_case_ ( _SCREAMING_SNAKE_CASE ): if "encoder." in name: __lowercase = name.replace("encoder." , "" ) if "cls_token" in name: __lowercase = name.replace("cls_token" , "videomae.embeddings.cls_token" ) if "decoder_pos_embed" in name: __lowercase = name.replace("decoder_pos_embed" , "decoder.decoder_pos_embed" ) if "pos_embed" in name and "decoder" not in name: __lowercase = name.replace("pos_embed" , "videomae.embeddings.position_embeddings" ) if "patch_embed.proj" in name: __lowercase = name.replace("patch_embed.proj" , "videomae.embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: __lowercase = name.replace("patch_embed.norm" , "videomae.embeddings.norm" ) if "decoder.blocks" in name: __lowercase = name.replace("decoder.blocks" , "decoder.decoder_layers" ) if "blocks" in name: __lowercase = name.replace("blocks" , "videomae.encoder.layer" ) if "attn.proj" in name: __lowercase = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name and "bias" not in name: __lowercase = name.replace("attn" , "attention.self" ) if "attn" in name: __lowercase = name.replace("attn" , "attention.attention" ) if "norm1" in name: __lowercase = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: __lowercase = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: __lowercase = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __lowercase = name.replace("mlp.fc2" , "output.dense" ) if "decoder_embed" in name: __lowercase = name.replace("decoder_embed" , "decoder.decoder_embed" ) if "decoder_norm" in name: __lowercase = name.replace("decoder_norm" , "decoder.decoder_norm" ) if "decoder_pred" in name: __lowercase = name.replace("decoder_pred" , "decoder.decoder_pred" ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: __lowercase = name.replace("norm.weight" , "videomae.layernorm.weight" ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: __lowercase = name.replace("norm.bias" , "videomae.layernorm.bias" ) if "head" in name and "decoder" not in name: __lowercase = name.replace("head" , "classifier" ) return name def snake_case_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(_SCREAMING_SNAKE_CASE ) if key.startswith("encoder." ): __lowercase = key.replace("encoder." , "" ) if "qkv" in key: __lowercase = key.split("." ) if key.startswith("decoder.blocks" ): __lowercase = config.decoder_hidden_size __lowercase = int(key_split[2] ) __lowercase = "decoder.decoder_layers." if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[dim : dim * 2, :] __lowercase = val[-dim:, :] else: __lowercase = config.hidden_size __lowercase = int(key_split[1] ) __lowercase = "videomae.encoder.layer." if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[dim : dim * 2, :] __lowercase = val[-dim:, :] else: __lowercase = val return orig_state_dict def snake_case_ ( ): __lowercase = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename="eating_spaghetti.npy" , repo_type="dataset" ) __lowercase = np.load(_SCREAMING_SNAKE_CASE ) return list(_SCREAMING_SNAKE_CASE ) def snake_case_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowercase = get_videomae_config(_SCREAMING_SNAKE_CASE ) if "finetuned" in model_name: __lowercase = VideoMAEForVideoClassification(_SCREAMING_SNAKE_CASE ) else: __lowercase = VideoMAEForPreTraining(_SCREAMING_SNAKE_CASE ) # download original checkpoint, hosted on Google Drive __lowercase = "pytorch_model.bin" gdown.cached_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , quiet=_SCREAMING_SNAKE_CASE ) __lowercase = torch.load(_SCREAMING_SNAKE_CASE , map_location="cpu" ) if "model" in files: __lowercase = files["model"] else: __lowercase = files["module"] __lowercase = convert_state_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) model.eval() # verify model on basic input __lowercase = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) __lowercase = prepare_video() __lowercase = image_processor(_SCREAMING_SNAKE_CASE , return_tensors="pt" ) if "finetuned" not in model_name: __lowercase = hf_hub_download(repo_id="hf-internal-testing/bool-masked-pos" , filename="bool_masked_pos.pt" ) __lowercase = torch.load(_SCREAMING_SNAKE_CASE ) __lowercase = model(**_SCREAMING_SNAKE_CASE ) __lowercase = outputs.logits __lowercase = [ "videomae-small-finetuned-kinetics", "videomae-small-finetuned-ssv2", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) "videomae-base-short", "videomae-base-short-finetuned-kinetics", "videomae-base", "videomae-base-finetuned-kinetics", "videomae-large", "videomae-large-finetuned-kinetics", "videomae-huge-finetuned-kinetics", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) "videomae-base-short-ssv2", "videomae-base-short-finetuned-ssv2", "videomae-base-ssv2", "videomae-base-finetuned-ssv2", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": __lowercase = torch.Size([1, 4_0_0] ) __lowercase = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": __lowercase = torch.Size([1, 1_7_4] ) __lowercase = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one __lowercase = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": __lowercase = torch.Size([1, 4_0_0] ) __lowercase = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": __lowercase = torch.Size([1, 4_0_0] ) __lowercase = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": __lowercase = torch.Size([1, 4_0_0] ) __lowercase = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": __lowercase = torch.Size([1, 4_0_0] ) __lowercase = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": __lowercase = torch.Size([1, 1_7_4] ) __lowercase = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": __lowercase = torch.Size([1, 1_4_0_8, 1_5_3_6] ) __lowercase = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": __lowercase = torch.Size([1, 1_7_4] ) __lowercase = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(F"""Model name not supported. Should be one of {model_names}""" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) else: print("Logits:" , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) print("Logits ok!" ) # verify loss, if applicable if model_name == "videomae-base-short": __lowercase = outputs.loss assert torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-4 ) print("Loss ok!" ) if pytorch_dump_folder_path is not None: print(F"""Saving model and image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if push_to_hub: print("Pushing to the hub..." ) model.push_to_hub(_SCREAMING_SNAKE_CASE , organization="nielsr" ) if __name__ == "__main__": snake_case__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4""", type=str, help=( """URL of the original PyTorch checkpoint (on Google Drive) you'd like to convert. Should be a direct""" """ download link.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default="""/Users/nielsrogge/Documents/VideoMAE/Test""", type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--model_name""", default="""videomae-base""", type=str, help="""Name of the model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) snake_case__ : Optional[int] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
402
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' def lowerCamelCase__ ( A : str ): '''simple docstring''' return str(A ) == str(A )[::-1] def lowerCamelCase__ ( A : Optional[Any] ): '''simple docstring''' return int(A ) + int(str(A )[::-1] ) def lowerCamelCase__ ( A : Tuple = 1_00_00 ): '''simple docstring''' UpperCAmelCase = [] for num in range(1 , A ): UpperCAmelCase = 0 UpperCAmelCase = num while iterations < 50: UpperCAmelCase = sum_reverse(A ) iterations += 1 if is_palindrome(A ): break else: lychrel_nums.append(A ) return len(A ) if __name__ == "__main__": print(F"""{solution() = }""")
210
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
0
'''simple docstring''' # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path __lowerCAmelCase = Path(__file__).resolve().parents[3] / """src""" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) __lowerCAmelCase = {"""base""": """patrickvonplaten/wav2vec2_tiny_random""", """robust""": """patrickvonplaten/wav2vec2_tiny_random_robust"""} __lowerCAmelCase = """zero2""" __lowerCAmelCase = """zero3""" __lowerCAmelCase = [ZEROa, ZEROa] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = parameterized.to_safe_name("""_""".join(str(_SCREAMING_SNAKE_CASE ) for x in param.args ) ) return f"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test __lowerCAmelCase = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class _lowerCAmelCase ( _lowerCAmelCase ): '''simple docstring''' @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> Tuple: self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> str: self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) def lowercase (self , UpperCAmelCase ) -> Any: pass def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 10 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = True , ) -> str: _snake_case = models[model] _snake_case = self.run_trainer( stage=_lowerCAmelCase , model_name=_lowerCAmelCase , eval_steps=_lowerCAmelCase , num_train_epochs=1 , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) self.do_checks(_lowerCAmelCase ) return output_dir def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 10 , UpperCAmelCase = 1 , UpperCAmelCase = True , UpperCAmelCase = True , ) -> Dict: _snake_case = self.get_auto_remove_tmp_dir("""./xxx""" , after=_lowerCAmelCase ) _snake_case = f"""\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(_lowerCAmelCase )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n """.split() if fpaa: args.extend(["""--fp16"""] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files _snake_case = f"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() _snake_case = [f"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] _snake_case = self.get_launcher(_lowerCAmelCase ) _snake_case = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_lowerCAmelCase , env=self.get_env() ) return output_dir def lowercase (self , UpperCAmelCase=False ) -> Dict: _snake_case = min(2 , get_gpu_count() ) if distributed else 1 return f"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
585
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
0
from math import isclose, sqrt def _UpperCAmelCase ( UpperCamelCase: int , UpperCamelCase: Dict , UpperCamelCase: str ): """simple docstring""" __lowerCAmelCase = point_y / 4 / point_x __lowerCAmelCase = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) __lowerCAmelCase = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) __lowerCAmelCase = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 __lowerCAmelCase = outgoing_gradient**2 + 4 __lowerCAmelCase = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) __lowerCAmelCase = (point_y - outgoing_gradient * point_x) ** 2 - 1_0_0 __lowerCAmelCase = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) __lowerCAmelCase = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point __lowerCAmelCase = x_minus if isclose(UpperCamelCase , UpperCamelCase ) else x_plus __lowerCAmelCase = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def _UpperCAmelCase ( UpperCamelCase: Dict = 1.4 , UpperCamelCase: str = -9.6 ): """simple docstring""" __lowerCAmelCase = 0 __lowerCAmelCase = first_x_coord __lowerCAmelCase = first_y_coord __lowerCAmelCase = (10.1 - point_y) / (0.0 - point_x) while not (-0.01 <= point_x <= 0.01 and point_y > 0): __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = next_point(UpperCamelCase , UpperCamelCase , UpperCamelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(f'''{solution() = }''')
611
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
0
'''simple docstring''' from bisect import bisect from itertools import accumulate def __a ( _UpperCamelCase: Any , _UpperCamelCase: Optional[int] , _UpperCamelCase: List[Any] , _UpperCamelCase: List[str] ) -> Tuple: """simple docstring""" _snake_case = sorted(zip(_UpperCamelCase , _UpperCamelCase ) , key=lambda _UpperCamelCase : x[0] / x[1] , reverse=_UpperCamelCase ) _snake_case , _snake_case = [i[0] for i in r], [i[1] for i in r] _snake_case = list(accumulate(_UpperCamelCase ) ) _snake_case = bisect(_UpperCamelCase , _UpperCamelCase ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
185
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
0
"""simple docstring""" import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __a ( _lowerCAmelCase , _lowerCAmelCase ): SCREAMING_SNAKE_CASE__ : List[Any] = 1 @register_to_config def __init__( self , a__ = 10_00 , a__ = None ): self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution _lowerCamelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. _lowerCamelCase = 4 # running values _lowerCamelCase = [] def snake_case_ ( self , a__ , a__ = None ): _lowerCamelCase = num_inference_steps _lowerCamelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] _lowerCamelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: _lowerCamelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: _lowerCamelCase = torch.sin(steps * math.pi / 2 ) ** 2 _lowerCamelCase = (1.0 - self.betas**2) ** 0.5 _lowerCamelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] _lowerCamelCase = timesteps.to(_lowerCAmelCase ) _lowerCamelCase = [] def snake_case_ ( self , a__ , a__ , a__ , a__ = True , ): if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) _lowerCamelCase = (self.timesteps == timestep).nonzero().item() _lowerCamelCase = timestep_index + 1 _lowerCamelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: _lowerCamelCase = self.ets[-1] elif len(self.ets ) == 2: _lowerCamelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: _lowerCamelCase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: _lowerCamelCase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) _lowerCamelCase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def snake_case_ ( self , a__ , *a__ , **a__ ): return sample def snake_case_ ( self , a__ , a__ , a__ , a__ ): _lowerCamelCase = self.alphas[timestep_index] _lowerCamelCase = self.betas[timestep_index] _lowerCamelCase = self.alphas[prev_timestep_index] _lowerCamelCase = self.betas[prev_timestep_index] _lowerCamelCase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) _lowerCamelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ): return self.config.num_train_timesteps
650
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
0
import torch def SCREAMING_SNAKE_CASE__ ( ) -> str: if torch.cuda.is_available(): __lowerCamelCase : Optional[Any] = torch.cuda.device_count() else: __lowerCamelCase : Optional[Any] = 0 print(F"Successfully ran on {num_gpus} GPUs" ) if __name__ == "__main__": main()
652
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class UpperCAmelCase : A__ : List[str] = BlenderbotConfig A__ : int = {} A__ : Optional[Any] = 'gelu' def __init__( self : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[int]=7 , __lowerCamelCase : str=True , __lowerCamelCase : Any=False , __lowerCamelCase : List[str]=9_9 , __lowerCamelCase : List[Any]=3_2 , __lowerCamelCase : Any=2 , __lowerCamelCase : str=4 , __lowerCamelCase : Any=3_7 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=0.1 , __lowerCamelCase : Dict=2_0 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : int=1 , __lowerCamelCase : Optional[int]=0 , ): """simple docstring""" _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id def __UpperCAmelCase ( self : int ): """simple docstring""" _snake_case = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _snake_case = prepare_blenderbot_inputs_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return config, inputs_dict def __UpperCAmelCase ( self : List[str] , __lowerCamelCase : int , __lowerCamelCase : List[Any] ): """simple docstring""" _snake_case = TFBlenderbotModel(config=_lowerCAmelCase ).get_decoder() _snake_case = inputs_dict['''input_ids'''] _snake_case = input_ids[:1, :] _snake_case = inputs_dict['''attention_mask'''][:1, :] _snake_case = inputs_dict['''head_mask'''] _snake_case = 1 # first forward pass _snake_case = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , head_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase ) _snake_case , _snake_case = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] _snake_case = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case = output_from_no_past[:, -3:, random_slice_idx] _snake_case = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_lowerCAmelCase , _lowerCAmelCase , rtol=1E-3 ) def snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , ) -> Optional[int]: if attention_mask is None: _snake_case = tf.cast(tf.math.not_equal(lowerCAmelCase_ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class UpperCAmelCase ( _lowerCAmelCase,_lowerCAmelCase,unittest.TestCase ): A__ : Union[str, Any] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () A__ : List[Any] = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () A__ : Union[str, Any] = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) A__ : Any = True A__ : Dict = False A__ : Optional[int] = False def __UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" _snake_case = TFBlenderbotModelTester(self ) _snake_case = ConfigTester(self , config_class=_lowerCAmelCase ) def __UpperCAmelCase ( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Dict ): """simple docstring""" _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowerCAmelCase ) @require_tokenizers @require_tf class UpperCAmelCase ( unittest.TestCase ): A__ : str = ['My friends are cool but they eat too many carbs.'] A__ : List[str] = 'facebook/blenderbot-400M-distill' @cached_property def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def __UpperCAmelCase ( self : List[str] ): """simple docstring""" _snake_case = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def __UpperCAmelCase ( self : str ): """simple docstring""" _snake_case = self.tokenizer(self.src_text , return_tensors='''tf''' ) _snake_case = self.model.generate( model_inputs.input_ids , ) _snake_case = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_lowerCAmelCase )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
103
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
0
import math def __magic_name__ ( ) -> str: """simple docstring""" lowercase_ : int = input("""Enter message: """ ) lowercase_ : Optional[int] = int(input(f"""Enter key [2-{len(lowercase ) - 1}]: """ ) ) lowercase_ : Dict = input("""Encryption/Decryption [e/d]: """ ) if mode.lower().startswith("""e""" ): lowercase_ : Any = encrypt_message(lowercase , lowercase ) elif mode.lower().startswith("""d""" ): lowercase_ : str = decrypt_message(lowercase , lowercase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f"""Output:\n{text + "|"}""" ) def __magic_name__ ( lowercase , lowercase ) -> Optional[int]: """simple docstring""" lowercase_ : Union[str, Any] = [""""""] * key for col in range(lowercase ): lowercase_ : Any = col while pointer < len(lowercase ): cipher_text[col] += message[pointer] pointer += key return "".join(lowercase ) def __magic_name__ ( lowercase , lowercase ) -> Tuple: """simple docstring""" lowercase_ : List[str] = math.ceil(len(lowercase ) / key ) lowercase_ : str = key lowercase_ : Optional[int] = (num_cols * num_rows) - len(lowercase ) lowercase_ : Optional[int] = [""""""] * num_cols lowercase_ : List[str] = 0 lowercase_ : Dict = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): lowercase_ : Union[str, Any] = 0 row += 1 return "".join(lowercase ) if __name__ == "__main__": import doctest doctest.testmod() main()
458
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
0
from itertools import product def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : Optional[Any] = sides_number lowerCamelCase__ : Tuple = max_face_number * dice_number lowerCamelCase__ : Any = [0] * (max_total + 1) lowerCamelCase__ : Optional[Any] = 1 lowerCamelCase__ : List[str] = range(_UpperCAmelCase , max_face_number + 1 ) for dice_numbers in product(_UpperCAmelCase , repeat=_UpperCAmelCase ): lowerCamelCase__ : str = sum(_UpperCAmelCase ) totals_frequencies[total] += 1 return totals_frequencies def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = total_frequency_distribution( sides_number=4 , dice_number=9 ) lowerCamelCase__ : Dict = total_frequency_distribution( sides_number=6 , dice_number=6 ) lowerCamelCase__ : Union[str, Any] = 0 lowerCamelCase__ : Optional[int] = 9 lowerCamelCase__ : Union[str, Any] = 4 * 9 lowerCamelCase__ : Dict = 6 for peter_total in range(_UpperCAmelCase , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) lowerCamelCase__ : Any = (4**9) * (6**6) lowerCamelCase__ : List[Any] = peter_wins_count / total_games_number lowerCamelCase__ : Tuple = round(_UpperCAmelCase , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(F"""{solution() = }""")
295
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
0
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' @property def _lowerCAmelCase ( self ): torch.manual_seed(0 ) A : Union[str, Any] = UNetaDModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=3, out_channels=3, down_block_types=("""DownBlock2D""", """AttnDownBlock2D"""), up_block_types=("""AttnUpBlock2D""", """UpBlock2D"""), ) return model def _lowerCAmelCase ( self ): A : Tuple = self.dummy_uncond_unet A : List[str] = PNDMScheduler() A : Union[str, Any] = PNDMPipeline(unet=_lowerCAmelCase, scheduler=_lowerCAmelCase ) pndm.to(_lowerCAmelCase ) pndm.set_progress_bar_config(disable=_lowerCAmelCase ) A : int = torch.manual_seed(0 ) A : Optional[int] = pndm(generator=_lowerCAmelCase, num_inference_steps=20, output_type="""numpy""" ).images A : Optional[Any] = torch.manual_seed(0 ) A : Any = pndm(generator=_lowerCAmelCase, num_inference_steps=20, output_type="""numpy""", return_dict=_lowerCAmelCase )[0] A : Optional[Any] = image[0, -3:, -3:, -1] A : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) A : str = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ): A : Dict = """google/ddpm-cifar10-32""" A : Any = UNetaDModel.from_pretrained(_lowerCAmelCase ) A : Dict = PNDMScheduler() A : Optional[int] = PNDMPipeline(unet=_lowerCAmelCase, scheduler=_lowerCAmelCase ) pndm.to(_lowerCAmelCase ) pndm.set_progress_bar_config(disable=_lowerCAmelCase ) A : List[str] = torch.manual_seed(0 ) A : Tuple = pndm(generator=_lowerCAmelCase, output_type="""numpy""" ).images A : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) A : List[Any] = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
662
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
0
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP snake_case__ : List[Any] = False try: snake_case__ : int = _is_package_available("""google.colab""") except ModuleNotFoundError: pass @input.register class _A : '''simple docstring''' def __init__( self : Dict , lowerCamelCase : str = None , lowerCamelCase : list = [] ): '''simple docstring''' __lowercase = 0 __lowercase = choices __lowercase = prompt if sys.platform == "win32": __lowercase = "*" else: __lowercase = "➔ " def _snake_case ( self : List[str] , lowerCamelCase : Dict , lowerCamelCase : str = "" ): '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 32 , _lowerCAmelCase ) else: forceWrite(self.choices[index] , _lowerCAmelCase ) def _snake_case ( self : Dict , lowerCamelCase : int ): '''simple docstring''' if index == self.position: forceWrite(f""" {self.arrow_char} """ ) self.write_choice(_lowerCAmelCase ) else: forceWrite(f""" {self.choices[index]}""" ) reset_cursor() def _snake_case ( self : Optional[Any] , lowerCamelCase : Direction , lowerCamelCase : int = 1 ): '''simple docstring''' __lowercase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(_lowerCAmelCase ) move_cursor(_lowerCAmelCase , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["up"] ) def _snake_case ( self : Any ): '''simple docstring''' self.move_direction(Direction.UP ) @input.mark(KEYMAP["down"] ) def _snake_case ( self : Optional[Any] ): '''simple docstring''' self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["newline"] ) def _snake_case ( self : Tuple ): '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) return self.position @input.mark(KEYMAP["interrupt"] ) def _snake_case ( self : Optional[Any] ): '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(_lowerCAmelCase )] for number in range(10 )] ) def _snake_case ( self : Union[str, Any] ): '''simple docstring''' __lowercase = int(chr(self.current_selection ) ) __lowercase = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , _lowerCAmelCase ) else: return else: return def _snake_case ( self : Optional[int] , lowerCamelCase : int = 0 ): '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , "\n" ) if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n" ) else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n" ) __lowercase = default_choice for i in range(len(self.choices ) ): self.print_choice(_lowerCAmelCase ) forceWrite("\n" ) move_cursor(len(self.choices ) - self.position , "UP" ) with cursor.hide(): while True: if in_colab: try: __lowercase = int(builtins.input() ) except ValueError: __lowercase = default_choice else: __lowercase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , "UP" ) clear_line() self.write_choice(_lowerCAmelCase , "\n" ) return choice
402
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
0
'''simple docstring''' def lowerCamelCase__ ( A : Any = 2_00_00_00 ): '''simple docstring''' UpperCAmelCase = [0 for i in range(n + 1 )] UpperCAmelCase = 1 UpperCAmelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A ): UpperCAmelCase = 1 UpperCAmelCase = 0 for i in range(A ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F"""{solution() = }""")
210
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = len(_SCREAMING_SNAKE_CASE ) while cur > 1: # Find the maximum number in arr _snake_case = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _snake_case = arr[mi::-1] + arr[mi + 1 : len(_SCREAMING_SNAKE_CASE )] # Reverse whole list _snake_case = arr[cur - 1 :: -1] + arr[cur : len(_SCREAMING_SNAKE_CASE )] cur -= 1 return arr if __name__ == "__main__": __lowerCAmelCase = input('Enter numbers separated by a comma:\n').strip() __lowerCAmelCase = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
585
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
0
import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class a ( _lowerCAmelCase , unittest.TestCase ): # TODO: is there an appropriate internal test set? lowercase_ : Union[str, Any] = 'ssube/stable-diffusion-x4-upscaler-onnx' def UpperCAmelCase__ ( self : Optional[int] , snake_case__ : Dict=0 ): """simple docstring""" __lowerCAmelCase = floats_tensor((1, 3, 128, 128) , rng=random.Random(_lowerCAmelCase ) ) __lowerCAmelCase = torch.manual_seed(_lowerCAmelCase ) __lowerCAmelCase = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase__ ( self : int ): """simple docstring""" __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = self.get_dummy_inputs() __lowerCAmelCase = pipe(**_lowerCAmelCase ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) __lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = self.get_dummy_inputs() __lowerCAmelCase = pipe(**_lowerCAmelCase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.6_8_9_8_8_9_2, 0.5_9_2_4_0_5_5_6, 0.5_2_4_9_9_5_2_7, 0.5_8_8_6_6_2_1_5, 0.5_2_2_5_8_2_3_5, 0.5_2_5_7_2_7_1_5, 0.6_2_4_1_4_4_7_3, 0.6_1_7_4_3_8_7, 0.6_2_1_4_9_6_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) __lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = self.get_dummy_inputs() __lowerCAmelCase = pipe(**_lowerCAmelCase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.7_6_5_9_2_7_8, 0.7_6_4_3_7_6_6_4, 0.7_5_5_7_9_1_0_7, 0.7_6_9_1_1_1_6, 0.7_7_6_6_6_9_8_6, 0.7_7_2_7_6_7_2, 0.7_7_5_8_6_6_4, 0.7_8_1_2_2_2_6, 0.7_6_9_4_2_5_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCAmelCase__ ( self : str ): """simple docstring""" __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) __lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = self.get_dummy_inputs() __lowerCAmelCase = pipe(**_lowerCAmelCase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def UpperCAmelCase__ ( self : str ): """simple docstring""" __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) __lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = self.get_dummy_inputs() __lowerCAmelCase = pipe(**_lowerCAmelCase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.7_7_4_2_4_4_9_6, 0.7_7_3_6_0_1, 0.7_6_4_5_2_8_8, 0.7_7_6_9_5_9_8, 0.7_7_7_2_7_3_9, 0.7_7_3_8_6_8_8, 0.7_8_1_8_7_2_3_3, 0.7_7_8_7_9_5_8_4, 0.7_6_7_0_4_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class a ( unittest.TestCase ): @property def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __lowerCAmelCase = ort.SessionOptions() __lowerCAmelCase = False return options def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __lowerCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) __lowerCAmelCase = init_image.resize((128, 128) ) # using the PNDM scheduler by default __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = "A fantasy landscape, trending on artstation" __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pipe( prompt=_lowerCAmelCase , image=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=10 , generator=_lowerCAmelCase , output_type="np" , ) __lowerCAmelCase = output.images __lowerCAmelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array([0.4_8_8_3, 0.4_9_4_7, 0.4_9_8_0, 0.4_9_7_5, 0.4_9_8_2, 0.4_9_8_0, 0.5_0_0_0, 0.5_0_0_6, 0.4_9_7_2] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" __lowerCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) __lowerCAmelCase = init_image.resize((128, 128) ) __lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , subfolder="scheduler" ) __lowerCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , scheduler=_lowerCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowerCAmelCase = "A fantasy landscape, trending on artstation" __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pipe( prompt=_lowerCAmelCase , image=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=20 , generator=_lowerCAmelCase , output_type="np" , ) __lowerCAmelCase = output.images __lowerCAmelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) __lowerCAmelCase = np.array( [0.5_0_1_7_3_7_5_3, 0.5_0_2_2_3_3_5_6, 0.5_0_2_0_3_9, 0.5_0_2_3_3_0_3_6, 0.5_0_2_3_7_2_5, 0.5_0_2_2_6_0_1, 0.5_0_1_8_7_5_8, 0.5_0_2_3_4_0_8_5, 0.5_0_2_4_1_5_6_6] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
611
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
0
'''simple docstring''' import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) UpperCamelCase_ : Dict = [ """cross_validation.py""", """gradient_accumulation.py""", """local_sgd.py""", """multi_process_metrics.py""", """memory.py""", """automatic_gradient_accumulation.py""", """fsdp_with_peak_mem_tracking.py""", """deepspeed_with_config_support.py""", """megatron_lm_gpt_pretraining.py""", ] class _a ( unittest.TestCase ): def _lowercase ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ) -> int: _snake_case = None _snake_case = os.path.abspath(os.path.join("examples" ,"by_feature" ) ) _snake_case = os.path.abspath("examples" ) for item in os.listdir(_lowerCAmelCase ): if item not in EXCLUDE_EXAMPLES: _snake_case = os.path.join(_lowerCAmelCase ,_lowerCAmelCase ) if os.path.isfile(_lowerCAmelCase ) and ".py" in item_path: with self.subTest( tested_script=_lowerCAmelCase ,feature_script=_lowerCAmelCase ,tested_section="main()" if parser_only else "training_function()" ,): _snake_case = compare_against_test( os.path.join(_lowerCAmelCase ,_lowerCAmelCase ) ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) _snake_case = "\n".join(_lowerCAmelCase ) if special_strings is not None: for string in special_strings: _snake_case = diff.replace(_lowerCAmelCase ,"" ) self.assertEqual(_lowerCAmelCase ,"" ) def _lowercase ( self ) -> List[Any]: self.one_complete_example("complete_nlp_example.py" ,_lowerCAmelCase ) self.one_complete_example("complete_nlp_example.py" ,_lowerCAmelCase ) def _lowercase ( self ) -> Tuple: _snake_case = os.path.abspath(os.path.join("examples" ,"cv_example.py" ) ) _snake_case = [ " " * 16 + "{\n\n", " " * 20 + "\"accuracy\": eval_metric[\"accuracy\"],\n\n", " " * 20 + "\"f1\": eval_metric[\"f1\"],\n\n", " " * 20 + "\"train_loss\": total_loss.item() / len(train_dataloader),\n\n", " " * 20 + "\"epoch\": epoch,\n\n", " " * 16 + "},\n\n", " " * 16 + "step=epoch,\n", " " * 12, " " * 8 + "for step, batch in enumerate(active_dataloader):\n", ] self.one_complete_example("complete_cv_example.py" ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) self.one_complete_example("complete_cv_example.py" ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) @mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """1"""} ) class _a ( _lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : List[Any] = False @classmethod def _lowercase ( cls ) -> Union[str, Any]: super().setUpClass() _snake_case = tempfile.mkdtemp() _snake_case = os.path.join(cls._tmpdir ,"default_config.yml" ) write_basic_config(save_location=cls.configPath ) _snake_case = ["accelerate", "launch", "--config_file", cls.configPath] @classmethod def _lowercase ( cls ) -> Tuple: super().tearDownClass() shutil.rmtree(cls._tmpdir ) def _lowercase ( self ) -> Optional[Any]: _snake_case = f"""\n examples/by_feature/checkpointing.py\n --checkpointing_steps epoch\n --output_dir {self.tmpdir}\n """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir ,"epoch_0" ) ) ) def _lowercase ( self ) -> Optional[int]: _snake_case = f"""\n examples/by_feature/checkpointing.py\n --checkpointing_steps 1\n --output_dir {self.tmpdir}\n """.split() _snake_case = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir ,"step_2" ) ) ) def _lowercase ( self ) -> Dict: _snake_case = f"""\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir ,'epoch_0' )}\n """.split() _snake_case = run_command(self._launch_args + testargs ,return_stdout=_lowerCAmelCase ) self.assertNotIn("epoch 0:" ,_lowerCAmelCase ) self.assertIn("epoch 1:" ,_lowerCAmelCase ) def _lowercase ( self ) -> Union[str, Any]: _snake_case = f"""\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir ,'step_2' )}\n """.split() _snake_case = run_command(self._launch_args + testargs ,return_stdout=_lowerCAmelCase ) if torch.cuda.is_available(): _snake_case = torch.cuda.device_count() else: _snake_case = 1 if num_processes > 1: self.assertNotIn("epoch 0:" ,_lowerCAmelCase ) self.assertIn("epoch 1:" ,_lowerCAmelCase ) else: self.assertIn("epoch 0:" ,_lowerCAmelCase ) self.assertIn("epoch 1:" ,_lowerCAmelCase ) @slow def _lowercase ( self ) -> Dict: _snake_case = "\n examples/by_feature/cross_validation.py\n --num_folds 2\n ".split() with mock.patch.dict(os.environ ,{"TESTING_MOCKED_DATALOADERS": "0"} ): _snake_case = run_command(self._launch_args + testargs ,return_stdout=_lowerCAmelCase ) _snake_case = re.findall("({.+})" ,_lowerCAmelCase ) _snake_case = [r for r in results if "accuracy" in r][-1] _snake_case = ast.literal_eval(_lowerCAmelCase ) self.assertGreaterEqual(results["accuracy"] ,0.7_5 ) def _lowercase ( self ) -> int: _snake_case = ["examples/by_feature/multi_process_metrics.py"] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ ,{"WANDB_MODE": "offline"} ) def _lowercase ( self ) -> str: with tempfile.TemporaryDirectory() as tmpdir: _snake_case = f"""\n examples/by_feature/tracking.py\n --with_tracking\n --project_dir {tmpdir}\n """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(_lowerCAmelCase ,"tracking" ) ) ) def _lowercase ( self ) -> List[Any]: _snake_case = ["examples/by_feature/gradient_accumulation.py"] run_command(self._launch_args + testargs ) def _lowercase ( self ) -> Optional[int]: _snake_case = ["examples/by_feature/local_sgd.py"] run_command(self._launch_args + testargs )
185
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
0
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin A_ : Optional[Any] =get_tests_dir("""fixtures/test_sentencepiece_bpe_char.model""") @require_sentencepiece @require_tokenizers class __a ( _lowerCAmelCase , unittest.TestCase ): SCREAMING_SNAKE_CASE__ : Tuple = SpeechTaTokenizer SCREAMING_SNAKE_CASE__ : int = False SCREAMING_SNAKE_CASE__ : Dict = True def snake_case_ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = SpeechTaTokenizer(_lowerCAmelCase ) _lowerCamelCase = AddedToken('<mask>' , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) _lowerCamelCase = mask_token tokenizer.add_special_tokens({'mask_token': mask_token} ) tokenizer.add_tokens(['<ctc_blank>'] ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case_ ( self , a__ ): _lowerCamelCase = 'this is a test' _lowerCamelCase = 'this is a test' return input_text, output_text def snake_case_ ( self , a__ , a__=False , a__=20 , a__=5 ): _lowerCamelCase , _lowerCamelCase = self.get_input_output_texts(_lowerCAmelCase ) _lowerCamelCase = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) _lowerCamelCase = tokenizer.decode(_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) return text, ids def snake_case_ ( self ): _lowerCamelCase = '<pad>' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def snake_case_ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-4] , 'œ' ) self.assertEqual(vocab_keys[-2] , '<mask>' ) self.assertEqual(vocab_keys[-1] , '<ctc_blank>' ) self.assertEqual(len(_lowerCAmelCase ) , 81 ) def snake_case_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def snake_case_ ( self ): _lowerCamelCase = self.get_tokenizers(do_lower_case=_lowerCAmelCase ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): _lowerCamelCase = tokenizer.vocab_size _lowerCamelCase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) _lowerCamelCase = ['aaaaa bbbbbb', 'cccccccccdddddddd'] _lowerCamelCase = tokenizer.add_tokens(_lowerCAmelCase ) _lowerCamelCase = tokenizer.vocab_size _lowerCamelCase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , len(_lowerCAmelCase ) ) self.assertEqual(_lowerCAmelCase , all_size + len(_lowerCAmelCase ) ) _lowerCamelCase = tokenizer.encode('aaaaa bbbbbb low cccccccccdddddddd l' , add_special_tokens=_lowerCAmelCase ) self.assertGreaterEqual(len(_lowerCAmelCase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) _lowerCamelCase = {'eos_token': '>>>>|||<||<<|<<', 'pad_token': '<<<<<|||>|>>>>|>'} _lowerCamelCase = tokenizer.add_special_tokens(_lowerCAmelCase ) _lowerCamelCase = tokenizer.vocab_size _lowerCamelCase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , len(_lowerCAmelCase ) ) self.assertEqual(_lowerCAmelCase , all_size_a + len(_lowerCAmelCase ) ) _lowerCamelCase = tokenizer.encode( '>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l' , add_special_tokens=_lowerCAmelCase ) self.assertGreaterEqual(len(_lowerCAmelCase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def snake_case_ ( self ): pass def snake_case_ ( self ): pass def snake_case_ ( self ): _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = tokenizer.tokenize('This is a test' ) # fmt: off self.assertListEqual(_lowerCAmelCase , [SPIECE_UNDERLINE, 'T', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'a', SPIECE_UNDERLINE, 't', 'e', 's', 't'] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) _lowerCamelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [SPIECE_UNDERLINE, 'I', SPIECE_UNDERLINE, 'w', 'a', 's', SPIECE_UNDERLINE, 'b', 'o', 'r', 'n', SPIECE_UNDERLINE, 'i', 'n', SPIECE_UNDERLINE, '92000', ',', SPIECE_UNDERLINE, 'a', 'n', 'd', SPIECE_UNDERLINE, 't', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'f', 'a', 'l', 's', 'é', '.'] ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) # fmt: off self.assertListEqual(_lowerCAmelCase , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on _lowerCamelCase = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [SPIECE_UNDERLINE, 'I', SPIECE_UNDERLINE, 'w', 'a', 's', SPIECE_UNDERLINE, 'b', 'o', 'r', 'n', SPIECE_UNDERLINE, 'i', 'n', SPIECE_UNDERLINE, '<unk>', ',', SPIECE_UNDERLINE, 'a', 'n', 'd', SPIECE_UNDERLINE, 't', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'f', 'a', 'l', 's', 'é', '.'] ) @slow def snake_case_ ( self ): _lowerCamelCase = [ 'Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides ' 'general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural ' 'Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained ' 'models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.', 'BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly ' 'conditioning on both left and right context in all layers.', 'The quick brown fox jumps over the lazy dog.', ] # fmt: off _lowerCamelCase = { 'input_ids': [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], 'attention_mask': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='microsoft/speecht5_asr' , revision='c5ef64c71905caeccde0e4462ef3f9077224c524' , sequences=_lowerCAmelCase , )
650
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
0
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> str: return "".join([hex(lowerCamelCase__ )[2:].zfill(2 ).upper() for byte in list(lowerCamelCase__ )] ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[Any]: if (len(lowerCamelCase__ ) % 2) != 0: raise ValueError( 'Base16 encoded data is invalid:\nData does not have an even number of hex digits.' ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(lowerCamelCase__ ) <= set('0123456789ABCDEF' ): raise ValueError( 'Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters.' ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 1_6 ) for i in range(0 , len(lowerCamelCase__ ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
652
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case = { """configuration_xlm_roberta_xl""": [ """XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaXLConfig""", """XLMRobertaXLOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaXLForCausalLM""", """XLMRobertaXLForMaskedLM""", """XLMRobertaXLForMultipleChoice""", """XLMRobertaXLForQuestionAnswering""", """XLMRobertaXLForSequenceClassification""", """XLMRobertaXLForTokenClassification""", """XLMRobertaXLModel""", """XLMRobertaXLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
103
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ) -> List[str]: """simple docstring""" lowercase_ : str = tempfile.mkdtemp() # fmt: off lowercase_ : Optional[Any] = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on lowercase_ : Optional[int] = dict(zip(_lowerCAmelCase, range(len(_lowerCAmelCase ) ) ) ) lowercase_ : str = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] lowercase_ : Union[str, Any] = {"""unk_token""": """<unk>"""} lowercase_ : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) lowercase_ : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowerCAmelCase ) ) lowercase_ : Any = { """do_resize""": True, """size""": 20, """do_center_crop""": True, """crop_size""": 18, """do_normalize""": True, """image_mean""": [0.48145466, 0.4578275, 0.40821073], """image_std""": [0.26862954, 0.26130258, 0.27577711], } lowercase_ : Optional[Any] = os.path.join(self.tmpdirname, _lowerCAmelCase ) with open(self.image_processor_file, """w""", encoding="""utf-8""" ) as fp: json.dump(_lowerCAmelCase, _lowerCAmelCase ) def snake_case__ ( self, **snake_case__ ) -> Optional[Any]: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname, pad_token="""!""", **_lowerCAmelCase ) def snake_case__ ( self, **snake_case__ ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname, pad_token="""!""", **_lowerCAmelCase ) def snake_case__ ( self, **snake_case__ ) -> List[str]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname, **_lowerCAmelCase ) def snake_case__ ( self ) -> Tuple: """simple docstring""" shutil.rmtree(self.tmpdirname ) def snake_case__ ( self ) -> Optional[int]: """simple docstring""" lowercase_ : int = [np.random.randint(2_55, size=(3, 30, 4_00), dtype=np.uinta )] lowercase_ : Tuple = [Image.fromarray(np.moveaxis(_lowerCAmelCase, 0, -1 ) ) for x in image_inputs] return image_inputs def snake_case__ ( self ) -> Tuple: """simple docstring""" lowercase_ : Any = self.get_tokenizer() lowercase_ : Optional[Any] = self.get_rust_tokenizer() lowercase_ : Union[str, Any] = self.get_image_processor() lowercase_ : List[Any] = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) processor_slow.save_pretrained(self.tmpdirname ) lowercase_ : List[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname, use_fast=_lowerCAmelCase ) lowercase_ : List[Any] = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) processor_fast.save_pretrained(self.tmpdirname ) lowercase_ : List[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab(), tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab(), tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab(), tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer, _lowerCAmelCase ) self.assertIsInstance(processor_fast.tokenizer, _lowerCAmelCase ) self.assertEqual(processor_slow.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor, _lowerCAmelCase ) self.assertIsInstance(processor_fast.image_processor, _lowerCAmelCase ) def snake_case__ ( self ) -> List[Any]: """simple docstring""" lowercase_ : Dict = OwlViTProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowercase_ : Optional[int] = self.get_tokenizer(bos_token="""(BOS)""", eos_token="""(EOS)""" ) lowercase_ : List[Any] = self.get_image_processor(do_normalize=_lowerCAmelCase ) lowercase_ : Optional[Any] = OwlViTProcessor.from_pretrained( self.tmpdirname, bos_token="""(BOS)""", eos_token="""(EOS)""", do_normalize=_lowerCAmelCase ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, _lowerCAmelCase ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, _lowerCAmelCase ) def snake_case__ ( self ) -> Union[str, Any]: """simple docstring""" lowercase_ : int = self.get_image_processor() lowercase_ : Optional[int] = self.get_tokenizer() lowercase_ : List[str] = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) lowercase_ : Optional[int] = self.prepare_image_inputs() lowercase_ : List[str] = image_processor(_lowerCAmelCase, return_tensors="""np""" ) lowercase_ : Dict = processor(images=_lowerCAmelCase, return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum(), input_processor[key].sum(), delta=1E-2 ) def snake_case__ ( self ) -> Optional[int]: """simple docstring""" lowercase_ : str = self.get_image_processor() lowercase_ : List[Any] = self.get_tokenizer() lowercase_ : List[str] = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) lowercase_ : Union[str, Any] = """lower newer""" lowercase_ : Union[str, Any] = processor(text=_lowerCAmelCase, return_tensors="""np""" ) lowercase_ : int = tokenizer(_lowerCAmelCase, return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist(), encoded_processor[key][0].tolist() ) def snake_case__ ( self ) -> int: """simple docstring""" lowercase_ : Any = self.get_image_processor() lowercase_ : Dict = self.get_tokenizer() lowercase_ : Tuple = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) lowercase_ : Dict = """lower newer""" lowercase_ : str = self.prepare_image_inputs() lowercase_ : Any = processor(text=_lowerCAmelCase, images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ), ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def snake_case__ ( self ) -> Any: """simple docstring""" lowercase_ : Any = """google/owlvit-base-patch32""" lowercase_ : str = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) lowercase_ : List[Any] = ["""cat""", """nasa badge"""] lowercase_ : Optional[Any] = processor(text=_lowerCAmelCase ) lowercase_ : int = 16 self.assertListEqual(list(inputs.keys() ), ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape, (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def snake_case__ ( self ) -> Optional[int]: """simple docstring""" lowercase_ : List[str] = """google/owlvit-base-patch32""" lowercase_ : int = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) lowercase_ : str = [["""cat""", """nasa badge"""], ["""person"""]] lowercase_ : str = processor(text=_lowerCAmelCase ) lowercase_ : List[Any] = 16 lowercase_ : Union[str, Any] = len(_lowerCAmelCase ) lowercase_ : str = max([len(_lowerCAmelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ), ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape, (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def snake_case__ ( self ) -> Any: """simple docstring""" lowercase_ : Optional[Any] = """google/owlvit-base-patch32""" lowercase_ : int = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) lowercase_ : List[Any] = ["""cat""", """nasa badge"""] lowercase_ : Optional[int] = processor(text=_lowerCAmelCase ) lowercase_ : int = 16 lowercase_ : Any = inputs["""input_ids"""] lowercase_ : int = [ [4_94_06, 23_68, 4_94_07, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_94_06, 68_41, 1_13_01, 4_94_07, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ), ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape, (2, seq_length) ) self.assertListEqual(list(input_ids[0] ), predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ), predicted_ids[1] ) def snake_case__ ( self ) -> List[Any]: """simple docstring""" lowercase_ : Dict = self.get_image_processor() lowercase_ : str = self.get_tokenizer() lowercase_ : str = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) lowercase_ : Dict = self.prepare_image_inputs() lowercase_ : Dict = self.prepare_image_inputs() lowercase_ : str = processor(images=_lowerCAmelCase, query_images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ), ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def snake_case__ ( self ) -> Tuple: """simple docstring""" lowercase_ : List[Any] = self.get_image_processor() lowercase_ : Tuple = self.get_tokenizer() lowercase_ : Tuple = OwlViTProcessor(tokenizer=_lowerCAmelCase, image_processor=_lowerCAmelCase ) lowercase_ : int = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowercase_ : Optional[int] = processor.batch_decode(_lowerCAmelCase ) lowercase_ : List[str] = tokenizer.batch_decode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase )
458
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
0
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 1000 ) -> Tuple: return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
295
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
662
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING snake_case__ : Optional[Any] = logging.get_logger(__name__) snake_case__ : int = { """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class _A ( _lowerCAmelCase ): '''simple docstring''' _snake_case : Any = 'deta' _snake_case : Union[str, Any] = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Optional[Any] , lowerCamelCase : Dict=None , lowerCamelCase : Optional[Any]=900 , lowerCamelCase : Optional[Any]=2_048 , lowerCamelCase : int=6 , lowerCamelCase : List[str]=2_048 , lowerCamelCase : List[Any]=8 , lowerCamelCase : str=6 , lowerCamelCase : Dict=1_024 , lowerCamelCase : str=8 , lowerCamelCase : str=0.0 , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[Any]="relu" , lowerCamelCase : Optional[Any]=256 , lowerCamelCase : str=0.1 , lowerCamelCase : Dict=0.0 , lowerCamelCase : Dict=0.0 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Optional[int]=1.0 , lowerCamelCase : Optional[Any]=True , lowerCamelCase : str=False , lowerCamelCase : Optional[int]="sine" , lowerCamelCase : int=5 , lowerCamelCase : int=4 , lowerCamelCase : List[str]=4 , lowerCamelCase : List[str]=True , lowerCamelCase : int=300 , lowerCamelCase : str=True , lowerCamelCase : str=True , lowerCamelCase : Any=1 , lowerCamelCase : Dict=5 , lowerCamelCase : Dict=2 , lowerCamelCase : Any=1 , lowerCamelCase : str=1 , lowerCamelCase : Union[str, Any]=5 , lowerCamelCase : Tuple=2 , lowerCamelCase : Tuple=0.1 , lowerCamelCase : Optional[Any]=0.25 , **lowerCamelCase : Any , ): '''simple docstring''' if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) __lowercase = CONFIG_MAPPING["resnet"](out_features=["stage2", "stage3", "stage4"] ) else: if isinstance(_lowerCAmelCase , _lowerCAmelCase ): __lowercase = backbone_config.pop("model_type" ) __lowercase = CONFIG_MAPPING[backbone_model_type] __lowercase = config_class.from_dict(_lowerCAmelCase ) __lowercase = backbone_config __lowercase = num_queries __lowercase = max_position_embeddings __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = init_xavier_std __lowercase = encoder_layerdrop __lowercase = auxiliary_loss __lowercase = position_embedding_type # deformable attributes __lowercase = num_feature_levels __lowercase = encoder_n_points __lowercase = decoder_n_points __lowercase = two_stage __lowercase = two_stage_num_proposals __lowercase = with_box_refine __lowercase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = mask_loss_coefficient __lowercase = dice_loss_coefficient __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient __lowercase = focal_alpha super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def _snake_case ( self : Optional[Any] ): '''simple docstring''' return self.encoder_attention_heads @property def _snake_case ( self : Dict ): '''simple docstring''' return self.d_model def _snake_case ( self : List[Any] ): '''simple docstring''' __lowercase = copy.deepcopy(self.__dict__ ) __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output
402
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' from __future__ import annotations import numpy as np def lowerCamelCase__ ( A : Any ): '''simple docstring''' return np.maximum(0 , A ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
210
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCAmelCase = { """configuration_mask2former""": [ """MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Mask2FormerConfig""", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ["""Mask2FormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ """MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """Mask2FormerForUniversalSegmentation""", """Mask2FormerModel""", """Mask2FormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
585
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
0
from math import sqrt def _UpperCAmelCase ( UpperCamelCase: List[Any] ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(sqrt(UpperCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _UpperCAmelCase ( UpperCamelCase: Optional[int] = 1_0_0_0_1 ): """simple docstring""" __lowerCAmelCase = 0 __lowerCAmelCase = 1 while count != nth and number < 3: number += 1 if is_prime(UpperCamelCase ): count += 1 while count != nth: number += 2 if is_prime(UpperCamelCase ): count += 1 return number if __name__ == "__main__": print(f'''{solution() = }''')
611
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
0
'''simple docstring''' import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase_ : int = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') class _a ( _lowerCAmelCase , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : List[Any] = BartphoTokenizer SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = True def _lowercase ( self ) -> List[str]: super().setUp() _snake_case = ["▁This", "▁is", "▁a", "▁t", "est"] _snake_case = dict(zip(_lowerCAmelCase ,range(len(_lowerCAmelCase ) ) ) ) _snake_case = {"unk_token": "<unk>"} _snake_case = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["monolingual_vocab_file"] ) with open(self.monolingual_vocab_file ,"w" ,encoding="utf-8" ) as fp: for token in vocab_tokens: fp.write(f"""{token} {vocab_tokens[token]}\n""" ) _snake_case = BartphoTokenizer(_lowerCAmelCase ,self.monolingual_vocab_file ,**self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self ,**_SCREAMING_SNAKE_CASE ) -> Optional[Any]: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname ,**_lowerCAmelCase ) def _lowercase ( self ,_SCREAMING_SNAKE_CASE ) -> Any: _snake_case = "This is a là test" _snake_case = "This is a<unk><unk> test" return input_text, output_text def _lowercase ( self ) -> str: _snake_case = BartphoTokenizer(_lowerCAmelCase ,self.monolingual_vocab_file ,**self.special_tokens_map ) _snake_case = "This is a là test" _snake_case = "▁This ▁is ▁a ▁l à ▁t est".split() _snake_case = tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase ,_lowerCAmelCase ) _snake_case = tokens + [tokenizer.unk_token] _snake_case = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) ,_lowerCAmelCase )
185
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
0
"""simple docstring""" def SCREAMING_SNAKE_CASE_ ( snake_case : List[Any] )-> Optional[int]: if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
650
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): _UpperCAmelCase : Union[str, Any] = StableDiffusionInpaintPipeline _UpperCAmelCase : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _UpperCAmelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _UpperCAmelCase : Any = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _UpperCAmelCase : Dict = frozenset([] ) def lowerCAmelCase ( self : Any): torch.manual_seed(0) __lowerCamelCase : str = UNetaDConditionModel( block_out_channels=(3_2, 6_4) ,layers_per_block=2 ,sample_size=3_2 ,in_channels=9 ,out_channels=4 ,down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') ,up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') ,cross_attention_dim=3_2 ,attention_head_dim=(2, 4) ,use_linear_projection=_lowerCAmelCase ,) __lowerCamelCase : Optional[int] = PNDMScheduler(skip_prk_steps=_lowerCAmelCase) torch.manual_seed(0) __lowerCamelCase : str = AutoencoderKL( block_out_channels=[3_2, 6_4] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=4 ,sample_size=1_2_8 ,) torch.manual_seed(0) __lowerCamelCase : Tuple = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=3_2 ,intermediate_size=3_7 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_0_0_0 ,hidden_act='gelu' ,projection_dim=5_1_2 ,) __lowerCamelCase : Optional[int] = CLIPTextModel(_lowerCAmelCase) __lowerCamelCase : int = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') __lowerCamelCase : List[str] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : Optional[int] ,SCREAMING_SNAKE_CASE__ : Optional[int]=0): __lowerCamelCase : Any = floats_tensor((1, 3, 3_2, 3_2) ,rng=random.Random(_lowerCAmelCase)).to(_lowerCAmelCase) __lowerCamelCase : Optional[int] = image.cpu().permute(0 ,2 ,3 ,1)[0] __lowerCamelCase : int = Image.fromarray(np.uinta(_lowerCAmelCase)).convert('RGB').resize((6_4, 6_4)) __lowerCamelCase : Any = Image.fromarray(np.uinta(image + 4)).convert('RGB').resize((6_4, 6_4)) if str(_lowerCAmelCase).startswith('mps'): __lowerCamelCase : Tuple = torch.manual_seed(_lowerCAmelCase) else: __lowerCamelCase : Optional[Any] = torch.Generator(device=_lowerCAmelCase).manual_seed(_lowerCAmelCase) __lowerCamelCase : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': init_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def lowerCAmelCase ( self : int): __lowerCamelCase : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Optional[Any] = self.get_dummy_components() __lowerCamelCase : Union[str, Any] = StableDiffusionInpaintPipeline(**_lowerCAmelCase) __lowerCamelCase : List[str] = sd_pipe.to(_lowerCAmelCase) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase) __lowerCamelCase : Any = self.get_dummy_inputs(_lowerCAmelCase) __lowerCamelCase : List[str] = sd_pipe(**_lowerCAmelCase).images __lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __lowerCamelCase : Union[str, Any] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def lowerCAmelCase ( self : Any): super().test_inference_batch_single_identical(expected_max_diff=3E-3) @slow @require_torch_gpu class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : Any): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase ( self : List[Any]): __lowerCamelCase : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png') __lowerCamelCase : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png') __lowerCamelCase : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench.npy') __lowerCamelCase : Dict = 'stabilityai/stable-diffusion-2-inpainting' __lowerCamelCase : str = StableDiffusionInpaintPipeline.from_pretrained(_lowerCAmelCase ,safety_checker=_lowerCAmelCase) pipe.to(_lowerCAmelCase) pipe.set_progress_bar_config(disable=_lowerCAmelCase) pipe.enable_attention_slicing() __lowerCamelCase : Union[str, Any] = 'Face of a yellow cat, high resolution, sitting on a park bench' __lowerCamelCase : Optional[int] = torch.manual_seed(0) __lowerCamelCase : str = pipe( prompt=_lowerCAmelCase ,image=_lowerCAmelCase ,mask_image=_lowerCAmelCase ,generator=_lowerCAmelCase ,output_type='np' ,) __lowerCamelCase : List[str] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 9E-3 def lowerCAmelCase ( self : int): __lowerCamelCase : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png') __lowerCamelCase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png') __lowerCamelCase : Optional[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench_fp16.npy') __lowerCamelCase : Optional[Any] = 'stabilityai/stable-diffusion-2-inpainting' __lowerCamelCase : str = StableDiffusionInpaintPipeline.from_pretrained( _lowerCAmelCase ,torch_dtype=torch.floataa ,safety_checker=_lowerCAmelCase ,) pipe.to(_lowerCAmelCase) pipe.set_progress_bar_config(disable=_lowerCAmelCase) pipe.enable_attention_slicing() __lowerCamelCase : List[str] = 'Face of a yellow cat, high resolution, sitting on a park bench' __lowerCamelCase : Optional[int] = torch.manual_seed(0) __lowerCamelCase : Any = pipe( prompt=_lowerCAmelCase ,image=_lowerCAmelCase ,mask_image=_lowerCAmelCase ,generator=_lowerCAmelCase ,output_type='np' ,) __lowerCamelCase : Optional[int] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 5E-1 def lowerCAmelCase ( self : Optional[Any]): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCamelCase : Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png') __lowerCamelCase : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png') __lowerCamelCase : List[str] = 'stabilityai/stable-diffusion-2-inpainting' __lowerCamelCase : Union[str, Any] = PNDMScheduler.from_pretrained(_lowerCAmelCase ,subfolder='scheduler') __lowerCamelCase : Union[str, Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowerCAmelCase ,safety_checker=_lowerCAmelCase ,scheduler=_lowerCAmelCase ,torch_dtype=torch.floataa ,) pipe.to(_lowerCAmelCase) pipe.set_progress_bar_config(disable=_lowerCAmelCase) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() __lowerCamelCase : Optional[int] = 'Face of a yellow cat, high resolution, sitting on a park bench' __lowerCamelCase : str = torch.manual_seed(0) __lowerCamelCase : Optional[int] = pipe( prompt=_lowerCAmelCase ,image=_lowerCAmelCase ,mask_image=_lowerCAmelCase ,generator=_lowerCAmelCase ,num_inference_steps=2 ,output_type='np' ,) __lowerCamelCase : List[Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
652
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
0
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers snake_case = [int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def snake_case ( ) -> Tuple: _snake_case = os.path.dirname(os.path.realpath(lowerCAmelCase_ ) ) _snake_case = os.path.join(lowerCAmelCase_ , '''words.txt''' ) _snake_case = '''''' with open(lowerCAmelCase_ ) as f: _snake_case = f.readline() _snake_case = [word.strip('''\"''' ) for word in words.strip('''\r\n''' ).split(''',''' )] _snake_case = [ word for word in [sum(ord(lowerCAmelCase_ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowerCAmelCase_ ) if __name__ == "__main__": print(solution())
103
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
0
import os import sys import unittest UpperCAmelCase_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) UpperCAmelCase_ = os.path.join("""tests""", """models""", """bert""", """test_modeling_bert.py""") UpperCAmelCase_ = os.path.join("""tests""", """models""", """blip""", """test_modeling_blip.py""") class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ) -> List[Any]: """simple docstring""" lowercase_ : int = get_test_to_tester_mapping(_lowerCAmelCase ) lowercase_ : Tuple = get_test_to_tester_mapping(_lowerCAmelCase ) lowercase_ : Optional[Any] = {"""BertModelTest""": """BertModelTester"""} lowercase_ : List[Any] = { """BlipModelTest""": """BlipModelTester""", """BlipTextImageModelTest""": """BlipTextImageModelsModelTester""", """BlipTextModelTest""": """BlipTextModelTester""", """BlipTextRetrievalModelTest""": """BlipTextRetrievalModelTester""", """BlipVQAModelTest""": """BlipVQAModelTester""", """BlipVisionModelTest""": """BlipVisionModelTester""", } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase ) def snake_case__ ( self ) -> str: """simple docstring""" lowercase_ : Tuple = get_model_to_test_mapping(_lowerCAmelCase ) lowercase_ : List[Any] = get_model_to_test_mapping(_lowerCAmelCase ) lowercase_ : List[str] = { """BertForMaskedLM""": ["""BertModelTest"""], """BertForMultipleChoice""": ["""BertModelTest"""], """BertForNextSentencePrediction""": ["""BertModelTest"""], """BertForPreTraining""": ["""BertModelTest"""], """BertForQuestionAnswering""": ["""BertModelTest"""], """BertForSequenceClassification""": ["""BertModelTest"""], """BertForTokenClassification""": ["""BertModelTest"""], """BertLMHeadModel""": ["""BertModelTest"""], """BertModel""": ["""BertModelTest"""], } lowercase_ : Union[str, Any] = { """BlipForConditionalGeneration""": ["""BlipTextImageModelTest"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTest"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTest"""], """BlipModel""": ["""BlipModelTest"""], """BlipTextModel""": ["""BlipTextModelTest"""], """BlipVisionModel""": ["""BlipVisionModelTest"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase ) def snake_case__ ( self ) -> int: """simple docstring""" lowercase_ : List[Any] = get_model_to_tester_mapping(_lowerCAmelCase ) lowercase_ : List[Any] = get_model_to_tester_mapping(_lowerCAmelCase ) lowercase_ : List[Any] = { """BertForMaskedLM""": ["""BertModelTester"""], """BertForMultipleChoice""": ["""BertModelTester"""], """BertForNextSentencePrediction""": ["""BertModelTester"""], """BertForPreTraining""": ["""BertModelTester"""], """BertForQuestionAnswering""": ["""BertModelTester"""], """BertForSequenceClassification""": ["""BertModelTester"""], """BertForTokenClassification""": ["""BertModelTester"""], """BertLMHeadModel""": ["""BertModelTester"""], """BertModel""": ["""BertModelTester"""], } lowercase_ : Optional[Any] = { """BlipForConditionalGeneration""": ["""BlipTextImageModelsModelTester"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTester"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTester"""], """BlipModel""": ["""BlipModelTester"""], """BlipTextModel""": ["""BlipTextModelTester"""], """BlipVisionModel""": ["""BlipVisionModelTester"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ), _lowerCAmelCase )
458
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_groupvit""": [ """GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GroupViTConfig""", """GroupViTOnnxConfig""", """GroupViTTextConfig""", """GroupViTVisionConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ """GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GroupViTModel""", """GroupViTPreTrainedModel""", """GroupViTTextModel""", """GroupViTVisionModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ """TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFGroupViTModel""", """TFGroupViTPreTrainedModel""", """TFGroupViTTextModel""", """TFGroupViTVisionModel""", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
295
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
0
import numpy as np import torch from imwatermark import WatermarkEncoder # Copied from https://github.com/Stability-AI/generative-models/blob/613af104c6b85184091d42d374fef420eddb356d/scripts/demo/streamlit_helpers.py#L66 SCREAMING_SNAKE_CASE_:List[str] = 0B101_100_111_110_110_010_010_000_011_110_111_011_000_110_011_110 # bin(x)[2:] gives bits of x as str, use int to convert them to 0/1 SCREAMING_SNAKE_CASE_:List[Any] = [int(bit) for bit in bin(WATERMARK_MESSAGE)[2:]] class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self ): A : Optional[int] = WATERMARK_BITS A : int = WatermarkEncoder() self.encoder.set_watermark("""bits""", self.watermark ) def _lowerCAmelCase ( self, lowerCamelCase__ ): if images.shape[-1] < 256: return images A : Optional[Any] = (255 * (images / 2 + 0.5)).cpu().permute(0, 2, 3, 1 ).float().numpy() A : Union[str, Any] = [self.encoder.encode(_lowerCAmelCase, """dwtDct""" ) for image in images] A : Optional[int] = torch.from_numpy(np.array(_lowerCAmelCase ) ).permute(0, 3, 1, 2 ) A : str = torch.clamp(2 * (images / 255 - 0.5), min=-1.0, max=1.0 ) return images
662
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case__ : Any = logging.get_logger(__name__) snake_case__ : Any = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _A ( _lowerCAmelCase , _lowerCAmelCase ): '''simple docstring''' _snake_case : List[Any] = 'convnextv2' def __init__( self : List[Any] , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=4 , lowerCamelCase : int=4 , lowerCamelCase : Dict=None , lowerCamelCase : Optional[Any]=None , lowerCamelCase : List[Any]="gelu" , lowerCamelCase : int=0.02 , lowerCamelCase : Any=1e-12 , lowerCamelCase : Union[str, Any]=0.0 , lowerCamelCase : str=224 , lowerCamelCase : List[Any]=None , lowerCamelCase : Any=None , **lowerCamelCase : Union[str, Any] , ): '''simple docstring''' super().__init__(**_lowerCAmelCase ) __lowercase = num_channels __lowercase = patch_size __lowercase = num_stages __lowercase = [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes __lowercase = [3, 3, 9, 3] if depths is None else depths __lowercase = hidden_act __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = drop_path_rate __lowercase = image_size __lowercase = ["stem"] + [f"""stage{idx}""" for idx in range(1 , len(self.depths ) + 1 )] __lowercase , __lowercase = get_aligned_output_features_output_indices( out_features=_lowerCAmelCase , out_indices=_lowerCAmelCase , stage_names=self.stage_names )
402
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
0
'''simple docstring''' def lowerCamelCase__ ( A : Tuple ): '''simple docstring''' for i in range(len(A ) - 1 , 0 , -1 ): UpperCAmelCase = False for j in range(A , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: UpperCAmelCase , UpperCAmelCase = unsorted[j - 1], unsorted[j] UpperCAmelCase = True for j in range(A ): if unsorted[j] > unsorted[j + 1]: UpperCAmelCase , UpperCAmelCase = unsorted[j + 1], unsorted[j] UpperCAmelCase = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() _lowercase : List[Any] = input("""Enter numbers separated by a comma:\n""").strip() _lowercase : int = [int(item) for item in user_input.split(""",""")] print(F"""{cocktail_shaker_sort(unsorted) = }""")
210
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' from functools import lru_cache def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = 2 _snake_case = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(_SCREAMING_SNAKE_CASE ) if n > 1: factors.add(_SCREAMING_SNAKE_CASE ) return factors @lru_cache def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): return len(unique_prime_factors(_SCREAMING_SNAKE_CASE ) ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): return len(set(_SCREAMING_SNAKE_CASE ) ) in (0, 1) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = 2 while True: # Increment each value of a generated range _snake_case = [base + i for i in range(_SCREAMING_SNAKE_CASE )] # Run elements through out unique_prime_factors function # Append our target number to the end. _snake_case = [upf_len(_SCREAMING_SNAKE_CASE ) for x in group] checker.append(_SCREAMING_SNAKE_CASE ) # If all numbers in the list are equal, return the group variable. if equality(_SCREAMING_SNAKE_CASE ): return group # Increment our base variable by 1 base += 1 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 4 ): _snake_case = run(_SCREAMING_SNAKE_CASE ) return results[0] if len(_SCREAMING_SNAKE_CASE ) else None if __name__ == "__main__": print(solution())
585
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
0
from __future__ import annotations from collections.abc import MutableSequence class a : def __init__( self : Optional[Any] , snake_case__ : int , snake_case__ : MutableSequence[float] ): """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( "The number of coefficients should be equal to the degree + 1." ) __lowerCAmelCase = list(_lowerCAmelCase ) __lowerCAmelCase = degree def __add__( self : Optional[int] , snake_case__ : Polynomial ): """simple docstring""" if self.degree > polynomial_a.degree: __lowerCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowerCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , snake_case__ : Polynomial ): """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ): """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , snake_case__ : Polynomial ): """simple docstring""" __lowerCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def UpperCAmelCase__ ( self : Optional[int] , snake_case__ : int | float ): """simple docstring""" __lowerCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ): """simple docstring""" __lowerCAmelCase = "" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ): """simple docstring""" return self.__str__() def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __lowerCAmelCase = [0] * self.degree for i in range(self.degree ): __lowerCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : int | float = 0 ): """simple docstring""" __lowerCAmelCase = [0] * (self.degree + 2) __lowerCAmelCase = constant for i in range(self.degree + 1 ): __lowerCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , snake_case__ : object ): """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , snake_case__ : object ): """simple docstring""" return not self.__eq__(_lowerCAmelCase )
611
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __a ( _UpperCamelCase: str ) -> Optional[Any]: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __a ( _UpperCamelCase: Optional[Any] ) -> List[Any]: """simple docstring""" _snake_case = create_tensor(_UpperCamelCase ) _snake_case = gather(_UpperCamelCase ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def __a ( _UpperCamelCase: Dict ) -> str: """simple docstring""" _snake_case = [state.process_index] _snake_case = gather_object(_UpperCamelCase ) assert len(_UpperCamelCase ) == state.num_processes, F"""{gathered_obj}, {len(_UpperCamelCase )} != {state.num_processes}""" assert gathered_obj == list(range(state.num_processes ) ), F"""{gathered_obj} != {list(range(state.num_processes ) )}""" def __a ( _UpperCamelCase: List[str] ) -> List[Any]: """simple docstring""" _snake_case = create_tensor(_UpperCamelCase ) _snake_case = broadcast(_UpperCamelCase ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def __a ( _UpperCamelCase: Tuple ) -> Optional[int]: """simple docstring""" if state.is_main_process: _snake_case = torch.arange(state.num_processes + 1 ).to(state.device ) else: _snake_case = torch.arange(state.num_processes ).to(state.device ) _snake_case = pad_across_processes(_UpperCamelCase ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def __a ( _UpperCamelCase: Optional[int] ) -> str: """simple docstring""" if state.num_processes != 2: return _snake_case = create_tensor(_UpperCamelCase ) _snake_case = reduce(_UpperCamelCase , "sum" ) _snake_case = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(_UpperCamelCase , _UpperCamelCase ), F"""{reduced_tensor} != {truth_tensor}""" def __a ( _UpperCamelCase: str ) -> Optional[Any]: """simple docstring""" if state.num_processes != 2: return _snake_case = create_tensor(_UpperCamelCase ) _snake_case = reduce(_UpperCamelCase , "mean" ) _snake_case = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(_UpperCamelCase , _UpperCamelCase ), F"""{reduced_tensor} != {truth_tensor}""" def __a ( _UpperCamelCase: str ) -> List[str]: """simple docstring""" main() def __a ( ) -> Optional[int]: """simple docstring""" _snake_case = PartialState() state.print(F"""State: {state}""" ) state.print("testing gather" ) test_gather(_UpperCamelCase ) state.print("testing gather_object" ) test_gather_object(_UpperCamelCase ) state.print("testing broadcast" ) test_broadcast(_UpperCamelCase ) state.print("testing pad_across_processes" ) test_pad_across_processes(_UpperCamelCase ) state.print("testing reduce_sum" ) test_reduce_sum(_UpperCamelCase ) state.print("testing reduce_mean" ) test_reduce_mean(_UpperCamelCase ) if __name__ == "__main__": main()
185
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
0
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch A_ : Dict =logging.get_logger(__name__) class __a : def __init__( self , a__ = None , a__ = None , a__=None , a__=None ): if not conversation_id: _lowerCamelCase = uuid.uuida() if past_user_inputs is None: _lowerCamelCase = [] if generated_responses is None: _lowerCamelCase = [] _lowerCamelCase = conversation_id _lowerCamelCase = past_user_inputs _lowerCamelCase = generated_responses _lowerCamelCase = text def __eq__( self , a__ ): if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def snake_case_ ( self , a__ , a__ = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) _lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: _lowerCamelCase = text def snake_case_ ( self ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _lowerCamelCase = None def snake_case_ ( self , a__ ): self.generated_responses.append(_lowerCAmelCase ) def snake_case_ ( self ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ): _lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _lowerCamelCase = 'user' if is_user else 'bot' output += F'{name} >> {text} \n' return output @add_end_docstrings( _lowerCAmelCase , R"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class __a ( _lowerCAmelCase ): def __init__( self , *a__ , **a__ ): super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) if self.tokenizer.pad_token_id is None: _lowerCamelCase = self.tokenizer.eos_token def snake_case_ ( self , a__=None , a__=None , a__=None , **a__ ): _lowerCamelCase = {} _lowerCamelCase = {} _lowerCamelCase = {} if min_length_for_response is not None: _lowerCamelCase = min_length_for_response if minimum_tokens is not None: _lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: _lowerCamelCase = generate_kwargs['max_length'] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(_lowerCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self , a__ , a__=0 , **a__ ): _lowerCamelCase = super().__call__(_lowerCAmelCase , num_workers=_lowerCAmelCase , **_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) and len(_lowerCAmelCase ) == 1: return outputs[0] return outputs def snake_case_ ( self , a__ , a__=32 ): if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise ValueError('ConversationalPipeline, expects Conversation as inputs' ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' 'Add user inputs with the conversation\'s `add_user_input` method' ) if hasattr(self.tokenizer , '_build_conversation_input_ids' ): _lowerCamelCase = self.tokenizer._build_conversation_input_ids(_lowerCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _lowerCamelCase = self._legacy_parse_and_tokenize(_lowerCAmelCase ) if self.framework == "pt": _lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": _lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def snake_case_ ( self , a__ , a__=10 , **a__ ): _lowerCamelCase = generate_kwargs.get('max_length' , self.model.config.max_length ) _lowerCamelCase = model_inputs['input_ids'].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) _lowerCamelCase = max_length - minimum_tokens _lowerCamelCase = model_inputs['input_ids'][:, -trim:] if "attention_mask" in model_inputs: _lowerCamelCase = model_inputs['attention_mask'][:, -trim:] _lowerCamelCase = model_inputs.pop('conversation' ) _lowerCamelCase = max_length _lowerCamelCase = self.model.generate(**_lowerCAmelCase , **_lowerCAmelCase ) if self.model.config.is_encoder_decoder: _lowerCamelCase = 1 else: _lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def snake_case_ ( self , a__ , a__=True ): _lowerCamelCase = model_outputs['output_ids'] _lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase , ) _lowerCamelCase = model_outputs['conversation'] conversation.mark_processed() conversation.append_response(_lowerCAmelCase ) return conversation def snake_case_ ( self , a__ ): _lowerCamelCase = self.tokenizer.eos_token_id _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) ) if len(_lowerCAmelCase ) > self.tokenizer.model_max_length: _lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
650
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
0
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class A_ ( _lowerCAmelCase ): def __init__( self : List[str] ,SCREAMING_SNAKE_CASE__ : pyspark.sql.DataFrame ,SCREAMING_SNAKE_CASE__ : Optional[NamedSplit] = None ,SCREAMING_SNAKE_CASE__ : Optional[Features] = None ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : str = None ,SCREAMING_SNAKE_CASE__ : bool = False ,SCREAMING_SNAKE_CASE__ : str = None ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : str = "arrow" ,**SCREAMING_SNAKE_CASE__ : List[str] ,): super().__init__( split=_lowerCAmelCase ,features=_lowerCAmelCase ,cache_dir=_lowerCAmelCase ,keep_in_memory=_lowerCAmelCase ,streaming=_lowerCAmelCase ,**_lowerCAmelCase ,) __lowerCamelCase : int = load_from_cache_file __lowerCamelCase : Optional[int] = file_format __lowerCamelCase : List[str] = Spark( df=_lowerCAmelCase ,features=_lowerCAmelCase ,cache_dir=_lowerCAmelCase ,working_dir=_lowerCAmelCase ,**_lowerCAmelCase ,) def lowerCAmelCase ( self : Dict): if self.streaming: return self.builder.as_streaming_dataset(split=self.split) __lowerCamelCase : int = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=_lowerCAmelCase ,file_format=self._file_format ,) return self.builder.as_dataset(split=self.split)
652
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class UpperCAmelCase ( unittest.TestCase ): def __UpperCAmelCase ( self : int ): """simple docstring""" _snake_case = torch.tensor([-1_0_0, -1, -0.1, 0, 0.1, 1.0, 1_0_0] ) _snake_case = get_activation('''gelu''' ) self.assertTrue(torch.allclose(gelu_python(_lowerCAmelCase ) , torch_builtin(_lowerCAmelCase ) ) ) self.assertFalse(torch.allclose(gelu_python(_lowerCAmelCase ) , gelu_new(_lowerCAmelCase ) ) ) def __UpperCAmelCase ( self : List[Any] ): """simple docstring""" _snake_case = torch.tensor([-1_0_0, -1, -0.1, 0, 0.1, 1.0, 1_0_0] ) _snake_case = get_activation('''gelu''' ) _snake_case = get_activation('''gelu_10''' ) _snake_case = torch_builtin(_lowerCAmelCase ) _snake_case = geluaa(_lowerCAmelCase ) _snake_case = torch.where(y_gelu_aa < 1_0.0 , 1 , 0 ) self.assertTrue(torch.max(_lowerCAmelCase ).item() == 1_0.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" get_activation('''gelu''' ) get_activation('''gelu_10''' ) get_activation('''gelu_fast''' ) get_activation('''gelu_new''' ) get_activation('''gelu_python''' ) get_activation('''gelu_pytorch_tanh''' ) get_activation('''linear''' ) get_activation('''mish''' ) get_activation('''quick_gelu''' ) get_activation('''relu''' ) get_activation('''sigmoid''' ) get_activation('''silu''' ) get_activation('''swish''' ) get_activation('''tanh''' ) with self.assertRaises(_lowerCAmelCase ): get_activation('''bogus''' ) with self.assertRaises(_lowerCAmelCase ): get_activation(_lowerCAmelCase ) def __UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" _snake_case = get_activation('''gelu''' ) _snake_case = 1 _snake_case = get_activation('''gelu''' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(_lowerCAmelCase ): _snake_case = acta.a
103
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
0
def __magic_name__ ( lowercase ) -> List[Any]: """simple docstring""" if collection == []: return [] # get some information about the collection lowercase_ : Dict = len(lowercase ) lowercase_ : Optional[int] = max(lowercase ) lowercase_ : Tuple = min(lowercase ) # create the counting array lowercase_ : int = coll_max + 1 - coll_min lowercase_ : Tuple = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowercase ): lowercase_ : List[str] = counting_arr[i] + counting_arr[i - 1] # create the output collection lowercase_ : List[Any] = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowercase ) ): lowercase_ : Optional[Any] = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def __magic_name__ ( lowercase ) -> Optional[Any]: """simple docstring""" return "".join([chr(lowercase ) for i in counting_sort([ord(lowercase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" UpperCAmelCase_ = input("""Enter numbers separated by a comma:\n""").strip() UpperCAmelCase_ = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
458
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
0
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder _UpperCAmelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name _UpperCAmelCase : List[Any] = 2_56 class lowerCAmelCase ( _lowerCAmelCase ): UpperCAmelCase__ = ['melgan'] def __init__( self : Dict , UpperCAmelCase : SpectrogramNotesEncoder , UpperCAmelCase : SpectrogramContEncoder , UpperCAmelCase : TaFilmDecoder , UpperCAmelCase : DDPMScheduler , UpperCAmelCase : OnnxRuntimeModel if is_onnx_available() else Any , ) -> None: super().__init__() # From MELGAN lowerCamelCase__ : Tuple = math.log(1e-5 ) # Matches MelGAN training. lowerCamelCase__ : List[Any] = 4.0 # Largest value for most examples lowerCamelCase__ : Any = 128 self.register_modules( notes_encoder=_lowerCAmelCase , continuous_encoder=_lowerCAmelCase , decoder=_lowerCAmelCase , scheduler=_lowerCAmelCase , melgan=_lowerCAmelCase , ) def A_ ( self : List[str] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=(-1.0, 1.0) , UpperCAmelCase : Optional[Any]=False ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = output_range if clip: lowerCamelCase__ : Optional[Any] = torch.clip(_lowerCAmelCase , self.min_value , self.max_value ) # Scale to [0, 1]. lowerCamelCase__ : int = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def A_ ( self : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[Any]=(-1.0, 1.0) , UpperCAmelCase : Tuple=False ) -> List[str]: lowerCamelCase__ , lowerCamelCase__ : int = input_range lowerCamelCase__ : List[str] = torch.clip(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if clip else outputs # Scale to [0, 1]. lowerCamelCase__ : List[str] = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def A_ ( self : str , UpperCAmelCase : str , UpperCAmelCase : int , UpperCAmelCase : str ) -> Optional[int]: lowerCamelCase__ : str = input_tokens > 0 lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.notes_encoder( encoder_input_tokens=_lowerCAmelCase , encoder_inputs_mask=_lowerCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : str = self.continuous_encoder( encoder_inputs=_lowerCAmelCase , encoder_inputs_mask=_lowerCAmelCase ) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def A_ ( self : Dict , UpperCAmelCase : Any , UpperCAmelCase : List[Any] , UpperCAmelCase : List[Any] ) -> Dict: lowerCamelCase__ : Optional[int] = noise_time if not torch.is_tensor(_lowerCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device ) elif torch.is_tensor(_lowerCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(input_tokens.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowerCamelCase__ : Optional[Any] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device ) lowerCamelCase__ : Union[str, Any] = self.decoder( encodings_and_masks=_lowerCAmelCase , decoder_input_tokens=_lowerCAmelCase , decoder_noise_time=_lowerCAmelCase ) return logits @torch.no_grad() def __call__( self : Tuple , UpperCAmelCase : List[List[int]] , UpperCAmelCase : Optional[torch.Generator] = None , UpperCAmelCase : int = 100 , UpperCAmelCase : bool = True , UpperCAmelCase : str = "numpy" , UpperCAmelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , UpperCAmelCase : int = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or callback_steps <= 0) ): raise ValueError( F"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" F""" {type(_lowerCAmelCase )}.""" ) lowerCamelCase__ : str = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa ) lowerCamelCase__ : int = np.zeros([1, 0, self.n_dims] , np.floataa ) lowerCamelCase__ : Dict = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=_lowerCAmelCase , device=self.device ) for i, encoder_input_tokens in enumerate(_lowerCAmelCase ): if i == 0: lowerCamelCase__ : Any = torch.from_numpy(pred_mel[:1].copy() ).to( device=self.device , dtype=self.decoder.dtype ) # The first chunk has no previous context. lowerCamelCase__ : List[str] = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=_lowerCAmelCase , device=self.device ) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. lowerCamelCase__ : Optional[int] = ones lowerCamelCase__ : Optional[Any] = self.scale_features( _lowerCAmelCase , output_range=[-1.0, 1.0] , clip=_lowerCAmelCase ) lowerCamelCase__ : Dict = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens] ).to(device=self.device ) , continuous_inputs=_lowerCAmelCase , continuous_mask=_lowerCAmelCase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop lowerCamelCase__ : List[str] = randn_tensor( shape=encoder_continuous_inputs.shape , generator=_lowerCAmelCase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(_lowerCAmelCase ) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): lowerCamelCase__ : str = self.decode( encodings_and_masks=_lowerCAmelCase , input_tokens=_lowerCAmelCase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 lowerCamelCase__ : List[Any] = self.scheduler.step(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , generator=_lowerCAmelCase ).prev_sample lowerCamelCase__ : List[str] = self.scale_to_features(_lowerCAmelCase , input_range=[-1.0, 1.0] ) lowerCamelCase__ : List[Any] = mel[:1] lowerCamelCase__ : List[Any] = mel.cpu().float().numpy() lowerCamelCase__ : Dict = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1 ) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_lowerCAmelCase , _lowerCAmelCase ) logger.info('Generated segment' , _lowerCAmelCase ) if output_type == "numpy" and not is_onnx_available(): raise ValueError( 'Cannot return output in \'np\' format if ONNX is not available. Make sure to have ONNX installed or set \'output_type\' to \'mel\'.' ) elif output_type == "numpy" and self.melgan is None: raise ValueError( 'Cannot return output in \'np\' format if melgan component is not defined. Make sure to define `self.melgan` or set \'output_type\' to \'mel\'.' ) if output_type == "numpy": lowerCamelCase__ : List[Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa ) ) else: lowerCamelCase__ : List[Any] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=_lowerCAmelCase )
295
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
0
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever SCREAMING_SNAKE_CASE_:Union[str, Any] = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE__ ( _lowerCAmelCase ): '''simple docstring''' def __init__( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__=None ): super().__init__( _lowerCAmelCase, question_encoder_tokenizer=_lowerCAmelCase, generator_tokenizer=_lowerCAmelCase, index=_lowerCAmelCase, init_retrieval=_lowerCAmelCase, ) A : Optional[Any] = None def _lowerCAmelCase ( self, lowerCamelCase__ ): logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually A : Optional[int] = self._infer_socket_ifname() # avoid clash with the NCCL port A : List[str] = str(distributed_port + 1 ) A : List[str] = dist.new_group(ranks=_lowerCAmelCase, backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _lowerCAmelCase ( self ): return dist.get_rank(group=self.process_group ) == 0 def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__=torch.floataa ): A : Tuple = torch.empty(_lowerCAmelCase, dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase, src=0, scatter_list=_lowerCAmelCase, group=self.process_group ) return target_tensor def _lowerCAmelCase ( self ): A : str = psutil.net_if_addrs() # a hacky way to deal with varying network interface names A : Union[str, Any] = next((addr for addr in addrs if addr.startswith("""e""" )), _lowerCAmelCase ) return ifname def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__ ): if not dist.is_initialized(): A , A : int = self._main_retrieve(_lowerCAmelCase, _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training A : Dict = dist.get_world_size(group=self.process_group ) # gather logic A : Union[str, Any] = None if self._is_main(): A : Dict = [torch.empty(question_hidden_states.shape, dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ), dst=0, gather_list=_lowerCAmelCase, group=self.process_group ) # scatter logic A : str = question_hidden_states.shape[0] A : List[str] = [] A : Optional[Any] = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size A , A : Union[str, Any] = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy(), _lowerCAmelCase ) A , A : Dict = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) A : List[str] = self._chunk_tensor(_lowerCAmelCase, _lowerCAmelCase ) A : List[Any] = self._chunk_tensor(_lowerCAmelCase, _lowerCAmelCase ) A : Optional[int] = self._scattered(_lowerCAmelCase, [n_queries, n_docs], target_type=torch.intaa ) A : Dict = self._scattered(_lowerCAmelCase, [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
662
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
0
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument snake_case__ : Optional[int] = { """/attention/""": """/0/SelfAttention/""", """/self_attention/""": """/0/SelfAttention/""", """/encoder_decoder_attention/""": """/1/EncDecAttention/""", """value""": """v""", """query""": """q""", """key""": """k""", """out""": """o""", """pre_self_attention_layer_norm""": """0/layer_norm""", """pre_cross_attention_layer_norm""": """1/layer_norm""", """pre_attention_layer_norm""": """0/layer_norm""", # previously 1, but seems wrong """token_embedder""": """shared""", """encoder_norm""": """final_layer_norm""", """decoder_norm""": """final_layer_norm""", """relpos_bias/rel_embedding""": """block/0/layer/0/SelfAttention/relative_attention_bias/weight""", """router/router_weights/w/""": """router/classifier/""", """roer/roer_weights/w/""": """router/classifier/""", """logits_dense""": """lm_head""", } def snake_case_ ( _SCREAMING_SNAKE_CASE ): __lowercase = list(s_dict.keys() ) for key in keys: __lowercase = R".*/layers_(\d+)" __lowercase = key if re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowercase = re.sub(R"layers_(\d+)" , R"block/\1/layer" , _SCREAMING_SNAKE_CASE ) __lowercase = R"(encoder|decoder)\/" if re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowercase = re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).groups() if groups[0] == "encoder": __lowercase = re.sub(R"/mlp/" , R"/1/mlp/" , _SCREAMING_SNAKE_CASE ) __lowercase = re.sub(R"/pre_mlp_layer_norm/" , R"/1/layer_norm/" , _SCREAMING_SNAKE_CASE ) elif groups[0] == "decoder": __lowercase = re.sub(R"/mlp/" , R"/2/mlp/" , _SCREAMING_SNAKE_CASE ) __lowercase = re.sub(R"/pre_mlp_layer_norm/" , R"/2/layer_norm/" , _SCREAMING_SNAKE_CASE ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: __lowercase = new_key.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(F"""{key} -> {new_key}""" ) __lowercase = s_dict.pop(_SCREAMING_SNAKE_CASE ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __lowercase = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __lowercase = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: __lowercase = s_dict[key].shape[0] __lowercase = s_dict[key] for idx in range(_SCREAMING_SNAKE_CASE ): __lowercase = expert_weihts[idx] print(F"""{key} -> {key.replace('expert/' , 'nested fstring' )}""" ) s_dict.pop(_SCREAMING_SNAKE_CASE ) return s_dict snake_case__ : int = { """NUM_ENCODER_LAYERS""": """num_layers""", """NUM_DECODER_LAYERS""": """num_decoder_layers""", """NUM_HEADS""": """num_heads""", """HEAD_DIM""": """d_kv""", """EMBED_DIM""": """d_model""", """MLP_DIM""": """d_ff""", """NUM_SELECTED_EXPERTS""": """num_selected_experts""", """NUM_ENCODER_SPARSE_LAYERS""": """num_sparse_encoder_layers""", """NUM_DECODER_SPARSE_LAYERS""": """num_sparse_decoder_layers""", """dense.MlpBlock.activations""": """feed_forward_proj""", } def snake_case_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): import regex as re with open(_SCREAMING_SNAKE_CASE , "r" ) as f: __lowercase = f.read() __lowercase = re.findall(R"(.*) = ([0-9.]*)" , _SCREAMING_SNAKE_CASE ) __lowercase = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": __lowercase = float(_SCREAMING_SNAKE_CASE ) if "." in value else int(_SCREAMING_SNAKE_CASE ) __lowercase = re.findall(R"(.*activations) = \(\'(.*)\',\)" , _SCREAMING_SNAKE_CASE )[0] __lowercase = str(activation[1] ) __lowercase = num_experts __lowercase = SwitchTransformersConfig(**_SCREAMING_SNAKE_CASE ) return config def snake_case_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE="./" , _SCREAMING_SNAKE_CASE=8 ): print(F"""Loading flax weights from : {flax_checkpoint_path}""" ) __lowercase = checkpoints.load_tax_checkpoint(_SCREAMING_SNAKE_CASE ) if gin_file is not None: __lowercase = convert_gin_to_config(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: __lowercase = SwitchTransformersConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) __lowercase = SwitchTransformersForConditionalGeneration(_SCREAMING_SNAKE_CASE ) __lowercase = flax_params["target"] __lowercase = flatten_dict(_SCREAMING_SNAKE_CASE , sep="/" ) __lowercase = rename_keys(_SCREAMING_SNAKE_CASE ) __lowercase = unflatten_dict(_SCREAMING_SNAKE_CASE , sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(F"""Save PyTorch model to {pytorch_dump_path}""" ) pt_model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": snake_case__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the""" """ model architecture. If not provided, a `gin_file` has to be provided.""" ), ) parser.add_argument( """--gin_file""", default=None, type=str, required=False, help="""Path to the gin config file. If not provided, a `config_file` has to be passed """, ) parser.add_argument( """--config_name""", default=None, type=str, required=False, help="""Config name of SwitchTransformers model.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output pytorch model.""" ) parser.add_argument("""--num_experts""", default=8, type=int, required=False, help="""Number of experts""") snake_case__ : int = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
402
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase : List[str] = logging.get_logger(__name__) _lowercase : Tuple = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class UpperCamelCase__( _lowerCAmelCase ): __magic_name__ : List[Any] = 'big_bird' def __init__( self : str , lowerCAmelCase : Any=50358 , lowerCAmelCase : Optional[int]=768 , lowerCAmelCase : Union[str, Any]=12 , lowerCAmelCase : Optional[int]=12 , lowerCAmelCase : Optional[Any]=3072 , lowerCAmelCase : List[str]="gelu_new" , lowerCAmelCase : Optional[Any]=0.1 , lowerCAmelCase : int=0.1 , lowerCAmelCase : Optional[int]=4096 , lowerCAmelCase : Any=2 , lowerCAmelCase : List[Any]=0.02 , lowerCAmelCase : List[Any]=1E-12 , lowerCAmelCase : List[Any]=True , lowerCAmelCase : Union[str, Any]=0 , lowerCAmelCase : Tuple=1 , lowerCAmelCase : Dict=2 , lowerCAmelCase : Optional[int]=66 , lowerCAmelCase : Dict="block_sparse" , lowerCAmelCase : Optional[int]=True , lowerCAmelCase : str=False , lowerCAmelCase : str=64 , lowerCAmelCase : Union[str, Any]=3 , lowerCAmelCase : Union[str, Any]=None , **lowerCAmelCase : List[Any] , )-> Union[str, Any]: """simple docstring""" super().__init__( pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , sep_token_id=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = type_vocab_size UpperCAmelCase = layer_norm_eps UpperCAmelCase = use_cache UpperCAmelCase = rescale_embeddings UpperCAmelCase = attention_type UpperCAmelCase = use_bias UpperCAmelCase = block_size UpperCAmelCase = num_random_blocks UpperCAmelCase = classifier_dropout class UpperCamelCase__( _lowerCAmelCase ): @property def a__( self : List[str] )-> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCAmelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
210
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
0
'''simple docstring''' class _lowerCAmelCase : # Public class to implement a graph '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = row _snake_case = col _snake_case = graph def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order _snake_case = [-1, 0, 1, -1, 1, -1, 0, 1] _snake_case = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , _lowerCAmelCase ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , _lowerCAmelCase ) def lowercase (self ) -> int: # And finally, count all islands. _snake_case = [[False for j in range(self.COL )] for i in range(self.ROW )] _snake_case = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) count += 1 return count
585
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
0
def _UpperCAmelCase ( UpperCamelCase: Optional[int] ): """simple docstring""" return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("Program to check whether a number is a Perfect number or not...") UpperCamelCase_ = int(input("Enter number: ").strip()) print(f'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
611
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
0
'''simple docstring''' from pathlib import Path import fire from tqdm import tqdm def __a ( _UpperCamelCase: List[Any]="ro" , _UpperCamelCase: Any="en" , _UpperCamelCase: Optional[Any]="wmt16" , _UpperCamelCase: Optional[int]=None ) -> List[Any]: """simple docstring""" try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("run pip install datasets" ) _snake_case = F"""{src_lang}-{tgt_lang}""" print(F"""Converting {dataset}-{pair}""" ) _snake_case = datasets.load_dataset(_UpperCamelCase , _UpperCamelCase ) if save_dir is None: _snake_case = F"""{dataset}-{pair}""" _snake_case = Path(_UpperCamelCase ) save_dir.mkdir(exist_ok=_UpperCamelCase ) for split in ds.keys(): print(F"""Splitting {split} with {ds[split].num_rows} records""" ) # to save to val.source, val.target like summary datasets _snake_case = "val" if split == "validation" else split _snake_case = save_dir.joinpath(F"""{fn}.source""" ) _snake_case = save_dir.joinpath(F"""{fn}.target""" ) _snake_case = src_path.open("w+" ) _snake_case = tgt_path.open("w+" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): _snake_case = x["translation"] src_fp.write(ex[src_lang] + "\n" ) tgt_fp.write(ex[tgt_lang] + "\n" ) print(F"""Saved {dataset} dataset to {save_dir}""" ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
185
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
0
"""simple docstring""" from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( snake_case : Any , snake_case : List[str] )-> Dict: if b == 0: return (1, 0) ((_lowerCamelCase) , (_lowerCamelCase)) = extended_euclid(snake_case , a % b ) _lowerCamelCase = a // b return (y, x - k * y) def SCREAMING_SNAKE_CASE_ ( snake_case : Union[str, Any] , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : int )-> int: ((_lowerCamelCase) , (_lowerCamelCase)) = extended_euclid(snake_case , snake_case ) _lowerCamelCase = na * na _lowerCamelCase = ra * x * na + ra * y * na return (n % m + m) % m def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[int] , snake_case : Optional[Any] )-> Tuple: ((_lowerCamelCase) , (_lowerCamelCase)) = extended_euclid(snake_case , snake_case ) if b < 0: _lowerCamelCase = (b % n + n) % n return b def SCREAMING_SNAKE_CASE_ ( snake_case : List[Any] , snake_case : int , snake_case : Union[str, Any] , snake_case : List[Any] )-> Dict: _lowerCamelCase , _lowerCamelCase = invert_modulo(snake_case , snake_case ), invert_modulo(snake_case , snake_case ) _lowerCamelCase = na * na _lowerCamelCase = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name="""chinese_remainder_theorem""", verbose=True) testmod(name="""chinese_remainder_theorem2""", verbose=True) testmod(name="""invert_modulo""", verbose=True) testmod(name="""extended_euclid""", verbose=True)
650
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
0
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class A_ ( _lowerCAmelCase ): def lowerCAmelCase ( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : float): return 0.0 def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Optional[Any]: __lowerCamelCase : Optional[Any] = min([-2_0, np.min(fft_results[1 : samplerate // 2 - 1] )] ) __lowerCamelCase : Dict = max([2_0, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> int: __lowerCamelCase : Tuple = 5_1_2 __lowerCamelCase : str = [1] + [0] * (size - 1) __lowerCamelCase : int = [filter_type.process(lowerCamelCase__ ) for item in inputs] __lowerCamelCase : str = [0] * (samplerate - size) # zero-padding outputs += filler __lowerCamelCase : Union[str, Any] = np.abs(np.fft.fft(lowerCamelCase__ ) ) __lowerCamelCase : List[Any] = 2_0 * np.logaa(lowerCamelCase__ ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds __lowerCamelCase : Tuple = get_bounds(lowerCamelCase__ , lowerCamelCase__ ) plt.ylim(max([-8_0, bounds[0]] ) , min([8_0, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(lowerCamelCase__ ) plt.show() def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Optional[int]: __lowerCamelCase : List[Any] = 5_1_2 __lowerCamelCase : Any = [1] + [0] * (size - 1) __lowerCamelCase : Union[str, Any] = [filter_type.process(lowerCamelCase__ ) for item in inputs] __lowerCamelCase : int = [0] * (samplerate - size) # zero-padding outputs += filler __lowerCamelCase : Optional[Any] = np.angle(np.fft.fft(lowerCamelCase__ ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(lowerCamelCase__ , -2 * pi ) ) plt.show()
652
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
0
"""simple docstring""" def snake_case ( lowerCAmelCase_ ) -> Dict: _snake_case = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(2_7)) print(perfect_cube(4))
103
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
458
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { """post_extract_proj""": """feature_projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.upsample.0""": """encoder.upsample.projection""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """layer_norm""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: for attribute in key.split('.' ): lowerCamelCase__ : List[str] = getattr(_UpperCAmelCase , _UpperCAmelCase ) if weight_type is not None: lowerCamelCase__ : Dict = getattr(_UpperCAmelCase , _UpperCAmelCase ).shape else: lowerCamelCase__ : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowerCamelCase__ : List[Any] = value elif weight_type == "weight_g": lowerCamelCase__ : Optional[int] = value elif weight_type == "weight_v": lowerCamelCase__ : int = value elif weight_type == "bias": lowerCamelCase__ : Dict = value else: lowerCamelCase__ : List[Any] = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : List[str] = [] lowerCamelCase__ : Any = fairseq_model.state_dict() lowerCamelCase__ : List[Any] = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): lowerCamelCase__ : str = False if "conv_layers" in name: load_conv_layer( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) lowerCamelCase__ : Any = True else: for key, mapped_key in MAPPING.items(): lowerCamelCase__ : Optional[Any] = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: lowerCamelCase__ : Optional[int] = True if "*" in mapped_key: lowerCamelCase__ : int = name.split(_UpperCAmelCase )[0].split('.' )[-2] lowerCamelCase__ : Dict = mapped_key.replace('*' , _UpperCAmelCase ) if "weight_g" in name: lowerCamelCase__ : Dict = 'weight_g' elif "weight_v" in name: lowerCamelCase__ : str = 'weight_v' elif "weight" in name: lowerCamelCase__ : Dict = 'weight' elif "bias" in name: lowerCamelCase__ : List[Any] = 'bias' else: lowerCamelCase__ : Optional[Any] = None set_recursively(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) continue if not is_used: unused_weights.append(_UpperCAmelCase ) logger.warning(F"""Unused weights: {unused_weights}""" ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : Optional[int] = full_name.split('conv_layers.' )[-1] lowerCamelCase__ : List[Any] = name.split('.' ) lowerCamelCase__ : Optional[Any] = int(items[0] ) lowerCamelCase__ : Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowerCamelCase__ : int = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowerCamelCase__ : Any = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) lowerCamelCase__ : str = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) lowerCamelCase__ : Dict = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ : List[str] = SEWConfig() if is_finetuned: lowerCamelCase__ : List[Any] = model.wav_encoder.wav_model.cfg else: lowerCamelCase__ : Dict = model.cfg lowerCamelCase__ : Any = fs_config.conv_bias lowerCamelCase__ : Optional[int] = eval(fs_config.conv_feature_layers ) lowerCamelCase__ : str = [x[0] for x in conv_layers] lowerCamelCase__ : List[Any] = [x[1] for x in conv_layers] lowerCamelCase__ : Any = [x[2] for x in conv_layers] lowerCamelCase__ : Union[str, Any] = 'gelu' lowerCamelCase__ : int = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' lowerCamelCase__ : Dict = 0.0 lowerCamelCase__ : Optional[int] = fs_config.activation_fn.name lowerCamelCase__ : Tuple = fs_config.encoder_embed_dim lowerCamelCase__ : Union[str, Any] = 0.02 lowerCamelCase__ : Optional[Any] = fs_config.encoder_ffn_embed_dim lowerCamelCase__ : int = 1e-5 lowerCamelCase__ : Optional[int] = fs_config.encoder_layerdrop lowerCamelCase__ : Tuple = fs_config.encoder_attention_heads lowerCamelCase__ : Union[str, Any] = fs_config.conv_pos_groups lowerCamelCase__ : Tuple = fs_config.conv_pos lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = fs_config.encoder_layers lowerCamelCase__ : int = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: lowerCamelCase__ : Dict = model.cfg lowerCamelCase__ : Optional[int] = fs_config.final_dropout lowerCamelCase__ : Optional[int] = fs_config.layerdrop lowerCamelCase__ : Optional[Any] = fs_config.activation_dropout lowerCamelCase__ : Tuple = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 lowerCamelCase__ : List[Any] = fs_config.attention_dropout lowerCamelCase__ : Optional[Any] = fs_config.dropout_input lowerCamelCase__ : Optional[int] = fs_config.dropout lowerCamelCase__ : Optional[int] = fs_config.mask_channel_length lowerCamelCase__ : int = fs_config.mask_channel_prob lowerCamelCase__ : Tuple = fs_config.mask_length lowerCamelCase__ : List[str] = fs_config.mask_prob lowerCamelCase__ : List[str] = 'Wav2Vec2FeatureExtractor' lowerCamelCase__ : str = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=True ) -> Any: if is_finetuned: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: lowerCamelCase__ : int = SEWConfig.from_pretrained(_UpperCAmelCase ) else: lowerCamelCase__ : int = convert_config(model[0] , _UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = model[0].eval() lowerCamelCase__ : Optional[int] = True if config.feat_extract_norm == 'layer' else False lowerCamelCase__ : List[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , ) if is_finetuned: if dict_path: lowerCamelCase__ : Optional[Any] = Dictionary.load(_UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCamelCase__ : int = target_dict.pad_index lowerCamelCase__ : List[str] = target_dict.bos_index lowerCamelCase__ : str = target_dict.pad_index lowerCamelCase__ : Any = target_dict.bos_index lowerCamelCase__ : Tuple = target_dict.eos_index lowerCamelCase__ : Optional[Any] = len(target_dict.symbols ) lowerCamelCase__ : Optional[Any] = os.path.join(_UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(_UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(_UpperCAmelCase ) ) return os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , _UpperCAmelCase ) lowerCamelCase__ : Optional[int] = WavaVecaCTCTokenizer( _UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=_UpperCAmelCase , ) lowerCamelCase__ : Optional[Any] = WavaVecaProcessor(feature_extractor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) lowerCamelCase__ : int = SEWForCTC(_UpperCAmelCase ) else: lowerCamelCase__ : Optional[Any] = SEWModel(_UpperCAmelCase ) feature_extractor.save_pretrained(_UpperCAmelCase ) recursively_load_weights(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) hf_model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--is_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
295
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
0
SCREAMING_SNAKE_CASE_:int = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ SCREAMING_SNAKE_CASE_:Optional[int] = [{"""type""": """code""", """content""": INSTALL_CONTENT}] SCREAMING_SNAKE_CASE_:Tuple = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
662
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
0
from __future__ import annotations from math import ceil, floor, sqrt def snake_case_ ( _SCREAMING_SNAKE_CASE = 2_0_0_0_0_0_0 ): __lowercase = [0] __lowercase = 4_2 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowercase = 0 # the area corresponding to the grid that gives the product closest to target __lowercase = 0 # an estimate of b, using the quadratic formula __lowercase = 4_2 # the largest integer less than b_estimate __lowercase = 4_2 # the largest integer less than b_estimate __lowercase = 4_2 # the triangle number corresponding to b_floor __lowercase = 4_2 # the triangle number corresponding to b_ceil __lowercase = 4_2 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowercase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowercase = floor(_SCREAMING_SNAKE_CASE ) __lowercase = ceil(_SCREAMING_SNAKE_CASE ) __lowercase = triangle_numbers[b_floor] __lowercase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowercase = triangle_b_first_guess * triangle_a __lowercase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowercase = triangle_b_second_guess * triangle_a __lowercase = idx_a * b_ceil return area if __name__ == "__main__": print(F'''{solution() = }''')
402
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
0
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class UpperCamelCase__: def __init__( self : List[str] , lowerCAmelCase : Any , lowerCAmelCase : Any=13 , lowerCAmelCase : str=7 , lowerCAmelCase : Tuple=True , lowerCAmelCase : Dict=True , lowerCAmelCase : Any=True , lowerCAmelCase : Union[str, Any]=True , lowerCAmelCase : int=99 , lowerCAmelCase : str=32 , lowerCAmelCase : Dict=2 , lowerCAmelCase : Union[str, Any]=4 , lowerCAmelCase : Any=37 , lowerCAmelCase : Union[str, Any]="gelu" , lowerCAmelCase : Union[str, Any]=0.1 , lowerCAmelCase : Tuple=0.1 , lowerCAmelCase : str=512 , lowerCAmelCase : Optional[int]=16 , lowerCAmelCase : Optional[int]=2 , lowerCAmelCase : Optional[Any]=0.02 , lowerCAmelCase : str=3 , lowerCAmelCase : Tuple=4 , lowerCAmelCase : List[str]=None , lowerCAmelCase : str=1000 , )-> Dict: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope UpperCAmelCase = range_bbox def a__( self : Optional[Any] )-> List[Any]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCAmelCase = bbox[i, j, 3] UpperCAmelCase = bbox[i, j, 1] UpperCAmelCase = t if bbox[i, j, 2] < bbox[i, j, 0]: UpperCAmelCase = bbox[i, j, 2] UpperCAmelCase = bbox[i, j, 0] UpperCAmelCase = t UpperCAmelCase = tf.convert_to_tensor(_lowerCAmelCase ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a__( self : List[str] , lowerCAmelCase : int , lowerCAmelCase : Tuple , lowerCAmelCase : List[str] , lowerCAmelCase : Optional[int] , lowerCAmelCase : Tuple , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Any , lowerCAmelCase : List[Any] )-> Tuple: """simple docstring""" UpperCAmelCase = TFLayoutLMModel(config=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , token_type_ids=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a__( self : Any , lowerCAmelCase : Tuple , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Dict , lowerCAmelCase : Optional[int] , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Dict , lowerCAmelCase : int , lowerCAmelCase : Union[str, Any] )-> Tuple: """simple docstring""" UpperCAmelCase = TFLayoutLMForMaskedLM(config=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__( self : Tuple , lowerCAmelCase : List[str] , lowerCAmelCase : Optional[Any] , lowerCAmelCase : List[str] , lowerCAmelCase : int , lowerCAmelCase : str , lowerCAmelCase : Tuple , lowerCAmelCase : Any , lowerCAmelCase : int )-> Optional[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = TFLayoutLMForSequenceClassification(config=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a__( self : List[Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : Any , lowerCAmelCase : Optional[int] , lowerCAmelCase : Any , lowerCAmelCase : Any , lowerCAmelCase : Dict )-> Dict: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = TFLayoutLMForTokenClassification(config=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a__( self : str , lowerCAmelCase : int , lowerCAmelCase : List[Any] , lowerCAmelCase : Tuple , lowerCAmelCase : Optional[Any] , lowerCAmelCase : int , lowerCAmelCase : Optional[Any] , lowerCAmelCase : List[str] , lowerCAmelCase : Optional[int] )-> Optional[int]: """simple docstring""" UpperCAmelCase = TFLayoutLMForQuestionAnswering(config=_lowerCAmelCase ) UpperCAmelCase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a__( self : str )-> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_tf class UpperCamelCase__( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __magic_name__ : List[Any] = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __magic_name__ : Union[str, Any] = ( { 'feature-extraction': TFLayoutLMModel, 'fill-mask': TFLayoutLMForMaskedLM, 'text-classification': TFLayoutLMForSequenceClassification, 'token-classification': TFLayoutLMForTokenClassification, 'zero-shot': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __magic_name__ : str = False __magic_name__ : Dict = True __magic_name__ : Tuple = 10 def a__( self : Dict )-> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFLayoutLMModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def a__( self : List[str] )-> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def a__( self : List[Any] )-> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def a__( self : int )-> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCAmelCase ) def a__( self : Any )-> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def a__( self : List[Any] )-> List[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) def a__( self : List[str] )-> Optional[int]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) @slow def a__( self : Any )-> Any: """simple docstring""" for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFLayoutLMModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip('''Onnx compliancy broke with TF 2.10''' ) def a__( self : str )-> Any: """simple docstring""" pass def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase = tf.convert_to_tensor([[1_01,10_19,10_14,10_16,10_37,1_28_49,47_47,10_04,1_42_46,22_78,54_39,45_24,50_02,29_30,21_93,29_30,43_41,32_08,10_05,10_55,21_71,28_48,1_13_00,35_31,1_02],[1_01,40_70,40_34,70_20,10_24,30_58,10_15,10_13,28_61,10_13,60_70,1_92_74,27_72,62_05,2_78_14,1_61_47,1_61_47,43_43,20_47,1_02_83,1_09_69,1_43_89,10_12,23_38,1_02]] ) # noqa: E231 UpperCAmelCase = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 UpperCAmelCase = tf.convert_to_tensor([[[0,0,0,0],[4_23,2_37,4_40,2_51],[4_27,2_72,4_41,2_87],[4_19,1_15,4_37,1_29],[9_61,8_85,9_92,9_12],[2_56,38,3_30,58],[2_56,38,3_30,58],[3_36,42,3_53,57],[3_60,39,4_01,56],[3_60,39,4_01,56],[4_11,39,4_71,59],[4_79,41,5_28,59],[5_33,39,6_30,60],[67,1_13,1_34,1_31],[1_41,1_15,2_09,1_32],[68,1_49,1_33,1_66],[1_41,1_49,1_87,1_64],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[2_95,1_48,3_49,1_65],[4_41,1_49,4_92,1_66],[4_97,1_49,5_46,1_64],[64,2_01,1_25,2_18],[10_00,10_00,10_00,10_00]],[[0,0,0,0],[6_62,1_50,7_54,1_66],[6_65,1_99,7_42,2_11],[5_19,2_13,5_54,2_28],[5_19,2_13,5_54,2_28],[1_34,4_33,1_87,4_54],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[3_14,4_69,3_76,4_82],[5_04,6_84,5_82,7_06],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[6_10,7_49,6_52,7_65],[1_30,6_59,1_68,6_72],[1_76,6_57,2_37,6_72],[2_38,6_57,3_12,6_72],[4_43,6_53,6_28,6_72],[4_43,6_53,6_28,6_72],[7_16,3_01,8_25,3_17],[10_00,10_00,10_00,10_00]]] ) # noqa: E231 UpperCAmelCase = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) UpperCAmelCase = tf.convert_to_tensor([[-1_00,10,10,10,9,1,-1_00,7,7,-1_00,7,7,4,2,5,2,8,8,-1_00,-1_00,5,0,3,2,-1_00],[-1_00,12,12,12,-1_00,12,10,-1_00,-1_00,-1_00,-1_00,10,12,9,-1_00,-1_00,-1_00,10,10,10,9,12,-1_00,10,-1_00]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class UpperCamelCase__( unittest.TestCase ): @slow def a__( self : str )-> str: """simple docstring""" UpperCAmelCase = TFLayoutLMModel.from_pretrained('''microsoft/layoutlm-base-uncased''' ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = prepare_layoutlm_batch_inputs() # forward pass UpperCAmelCase = model(input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) # test the sequence output on [0, :3, :3] UpperCAmelCase = tf.convert_to_tensor( [[0.1785, -0.1947, -0.0425], [-0.3254, -0.2807, 0.2553], [-0.5391, -0.3322, 0.3364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , _lowerCAmelCase , atol=1E-3 ) ) # test the pooled output on [1, :3] UpperCAmelCase = tf.convert_to_tensor([-0.6580, -0.0214, 0.8552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , _lowerCAmelCase , atol=1E-3 ) ) @slow def a__( self : int )-> Any: """simple docstring""" UpperCAmelCase = TFLayoutLMForSequenceClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=2 ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = prepare_layoutlm_batch_inputs() # forward pass UpperCAmelCase = model( input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar UpperCAmelCase = outputs.loss UpperCAmelCase = (2,) self.assertEqual(loss.shape , _lowerCAmelCase ) # test the shape of the logits UpperCAmelCase = outputs.logits UpperCAmelCase = (2, 2) self.assertEqual(logits.shape , _lowerCAmelCase ) @slow def a__( self : List[str] )-> int: """simple docstring""" UpperCAmelCase = TFLayoutLMForTokenClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=13 ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = prepare_layoutlm_batch_inputs() # forward pass UpperCAmelCase = model( input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) # test the shape of the logits UpperCAmelCase = outputs.logits UpperCAmelCase = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , _lowerCAmelCase ) @slow def a__( self : List[Any] )-> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFLayoutLMForQuestionAnswering.from_pretrained('''microsoft/layoutlm-base-uncased''' ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = prepare_layoutlm_batch_inputs() # forward pass UpperCAmelCase = model(input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) # test the shape of the logits UpperCAmelCase = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , _lowerCAmelCase ) self.assertEqual(outputs.end_logits.shape , _lowerCAmelCase )
210
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
0
'''simple docstring''' import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel __lowerCAmelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } __lowerCAmelCase = AutoFeatureExtractor.from_pretrained('laion/clap-htsat-unfused', truncation='rand_trunc') def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ): _snake_case, _snake_case = create_model( """HTSAT-tiny""" , """roberta""" , _SCREAMING_SNAKE_CASE , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_SCREAMING_SNAKE_CASE , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = {} _snake_case = R""".*sequential.(\d+).*""" _snake_case = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _snake_case = key.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): # replace sequential layers with list _snake_case = re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).group(1 ) _snake_case = key.replace(f"""sequential.{sequential_layer}.""" , f"""layers.{int(_SCREAMING_SNAKE_CASE )//3}.linear.""" ) elif re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = int(re.match(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... _snake_case = 1 if projecton_layer == 0 else 2 _snake_case = key.replace(f"""_projection.{projecton_layer}.""" , f"""_projection.linear{transformers_projection_layer}.""" ) if "audio" and "qkv" in key: # split qkv into query key and value _snake_case = value _snake_case = mixed_qkv.size(0 ) // 3 _snake_case = mixed_qkv[:qkv_dim] _snake_case = mixed_qkv[qkv_dim : qkv_dim * 2] _snake_case = mixed_qkv[qkv_dim * 2 :] _snake_case = query_layer _snake_case = key_layer _snake_case = value_layer else: _snake_case = value return model_state_dict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ): _snake_case, _snake_case = init_clap(_SCREAMING_SNAKE_CASE , enable_fusion=_SCREAMING_SNAKE_CASE ) clap_model.eval() _snake_case = clap_model.state_dict() _snake_case = rename_state_dict(_SCREAMING_SNAKE_CASE ) _snake_case = ClapConfig() _snake_case = enable_fusion _snake_case = ClapModel(_SCREAMING_SNAKE_CASE ) # ignore the spectrogram embedding layer model.load_state_dict(_SCREAMING_SNAKE_CASE , strict=_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) transformers_config.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument('--enable_fusion', action='store_true', help='Whether to enable fusion or not') __lowerCAmelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
585
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
0
from __future__ import annotations from fractions import Fraction def _UpperCAmelCase ( UpperCamelCase: Dict , UpperCamelCase: Optional[Any] ): """simple docstring""" return ( num != den and num % 1_0 == den // 1_0 and (num // 1_0) / (den % 1_0) == num / den ) def _UpperCAmelCase ( UpperCamelCase: List[Any] ): """simple docstring""" __lowerCAmelCase = [] __lowerCAmelCase = 1_1 __lowerCAmelCase = int("1" + "0" * digit_len ) for num in range(UpperCamelCase , UpperCamelCase ): while den <= 9_9: if (num != den) and (num % 1_0 == den // 1_0) and (den % 1_0 != 0): if is_digit_cancelling(UpperCamelCase , UpperCamelCase ): solutions.append(F"{num}/{den}" ) den += 1 num += 1 __lowerCAmelCase = 1_0 return solutions def _UpperCAmelCase ( UpperCamelCase: Optional[Any] = 2 ): """simple docstring""" __lowerCAmelCase = 1.0 for fraction in fraction_list(UpperCamelCase ): __lowerCAmelCase = Fraction(UpperCamelCase ) result *= frac.denominator / frac.numerator return int(UpperCamelCase ) if __name__ == "__main__": print(solution())
611
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
0
'''simple docstring''' # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def __a ( *_UpperCamelCase: Optional[int] ) -> str: """simple docstring""" with open(_UpperCamelCase , "r" ) as fh: fcntl.flock(_UpperCamelCase , fcntl.LOCK_EX ) try: print(*_UpperCamelCase ) finally: fcntl.flock(_UpperCamelCase , fcntl.LOCK_UN ) UpperCamelCase_ : List[Any] = int(os.environ['''LOCAL_RANK''']) torch.cuda.set_device(local_rank) UpperCamelCase_ : str = torch.device('''cuda''', local_rank) UpperCamelCase_ : str = socket.gethostname() UpperCamelCase_ : Optional[int] = F'[{hostname}-{local_rank}]' try: # test distributed dist.init_process_group('''nccl''') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank UpperCamelCase_ : Dict = dist.get_rank() UpperCamelCase_ : Optional[int] = dist.get_world_size() printflock(F'{gpu} is OK (global rank: {rank}/{world_size})') dist.barrier() if rank == 0: printflock(F'pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}') except Exception: printflock(F'{gpu} is broken') raise
185
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
0
"""simple docstring""" from __future__ import annotations class __a : def __init__( self , a__ ): _lowerCamelCase = data _lowerCamelCase = None _lowerCamelCase = None def SCREAMING_SNAKE_CASE_ ( snake_case : int )-> Optional[int]: # In Order traversal of the tree if tree: display(tree.left ) print(tree.data ) display(tree.right ) def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[int] )-> List[Any]: return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def SCREAMING_SNAKE_CASE_ ( snake_case : int )-> Any: if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def SCREAMING_SNAKE_CASE_ ( )-> Union[str, Any]: # Main function for testing. _lowerCamelCase = Node(1 ) _lowerCamelCase = Node(2 ) _lowerCamelCase = Node(3 ) _lowerCamelCase = Node(4 ) _lowerCamelCase = Node(5 ) _lowerCamelCase = Node(6 ) _lowerCamelCase = Node(7 ) _lowerCamelCase = Node(8 ) _lowerCamelCase = Node(9 ) print(is_full_binary_tree(snake_case ) ) print(depth_of_tree(snake_case ) ) print('Tree is: ' ) display(snake_case ) if __name__ == "__main__": main()
650
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
0