code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations import math def snake_case ( lowerCamelCase ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = str(lowerCamelCase ) __lowercase = [n] for i in range(1 , len(lowerCamelCase ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def snake_case ( lowerCamelCase ): '''simple docstring''' if len(str(lowerCamelCase ) ) > 3: if not is_prime(int(str(lowerCamelCase )[-3:] ) ) or not is_prime(int(str(lowerCamelCase )[:3] ) ): return False return True def snake_case ( lowerCamelCase = 11 ): '''simple docstring''' __lowercase = [] __lowercase = 13 while len(lowerCamelCase ) != count: if validate(lowerCamelCase ): __lowercase = list_truncated_nums(lowerCamelCase ) if all(is_prime(lowerCamelCase ) for i in list_nums ): list_truncated_primes.append(lowerCamelCase ) num += 2 return list_truncated_primes def snake_case ( ): '''simple docstring''' return sum(compute_truncated_primes(11 ) ) if __name__ == "__main__": print(F'''{sum(compute_truncated_primes(11)) = }''')
80
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
80
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
1
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) __UpperCamelCase : Tuple = """hf-internal-testing/tiny-random-bert""" __UpperCamelCase : str = os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") __UpperCamelCase : Optional[Any] = """9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[Any] ) -> str: """simple docstring""" __lowercase = cached_file(_lowerCAmelCase , _lowerCAmelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(_lowerCAmelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) ) with open(os.path.join(_lowerCAmelCase , """refs""" , """main""" ) ) as f: __lowercase = f.read() self.assertEqual(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """snapshots""" , _lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(os.path.isfile(_lowerCAmelCase ) ) # File is cached at the same place the second time. __lowercase = cached_file(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) # Using a specific revision to test the full commit hash. __lowercase = cached_file(_lowerCAmelCase , _lowerCAmelCase , revision="""9b8c223""" ) self.assertEqual(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """snapshots""" , _lowerCAmelCase , _lowerCAmelCase ) ) def _a ( self : Optional[int] ) -> Any: """simple docstring""" with self.assertRaisesRegex(_lowerCAmelCase , """is not a valid model identifier""" ): __lowercase = cached_file("""tiny-random-bert""" , _lowerCAmelCase ) with self.assertRaisesRegex(_lowerCAmelCase , """is not a valid git identifier""" ): __lowercase = cached_file(_lowerCAmelCase , _lowerCAmelCase , revision="""aaaa""" ) with self.assertRaisesRegex(_lowerCAmelCase , """does not appear to have a file named""" ): __lowercase = cached_file(_lowerCAmelCase , """conf""" ) def _a ( self : Tuple ) -> str: """simple docstring""" with self.assertRaisesRegex(_lowerCAmelCase , """does not appear to have a file named""" ): __lowercase = cached_file(_lowerCAmelCase , """conf""" ) with open(os.path.join(_lowerCAmelCase , """refs""" , """main""" ) ) as f: __lowercase = f.read() self.assertTrue(os.path.isfile(os.path.join(_lowerCAmelCase , """.no_exist""" , _lowerCAmelCase , """conf""" ) ) ) __lowercase = cached_file(_lowerCAmelCase , """conf""" , _raise_exceptions_for_missing_entries=_lowerCAmelCase ) self.assertIsNone(_lowerCAmelCase ) __lowercase = cached_file(_lowerCAmelCase , """conf""" , local_files_only=_lowerCAmelCase , _raise_exceptions_for_missing_entries=_lowerCAmelCase ) self.assertIsNone(_lowerCAmelCase ) __lowercase = mock.Mock() __lowercase = 500 __lowercase = {} __lowercase = HTTPError __lowercase = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=_lowerCAmelCase ) as mock_head: __lowercase = cached_file(_lowerCAmelCase , """conf""" , _raise_exceptions_for_connection_errors=_lowerCAmelCase ) self.assertIsNone(_lowerCAmelCase ) # This check we did call the fake head request mock_head.assert_called() def _a ( self : Optional[int] ) -> str: """simple docstring""" self.assertTrue(has_file("""hf-internal-testing/tiny-bert-pt-only""" , _lowerCAmelCase ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , _lowerCAmelCase ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , _lowerCAmelCase ) ) def _a ( self : int ) -> Tuple: """simple docstring""" self.assertIsNone(get_file_from_repo("""bert-base-cased""" , """ahah.txt""" ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(_lowerCAmelCase , """is not a valid model identifier""" ): get_file_from_repo("""bert-base-case""" , _lowerCAmelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(_lowerCAmelCase , """is not a valid git identifier""" ): get_file_from_repo("""bert-base-cased""" , _lowerCAmelCase , revision="""ahaha""" ) __lowercase = get_file_from_repo("""bert-base-cased""" , _lowerCAmelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. __lowercase = json.loads(open(_lowerCAmelCase , """r""" ).read() ) self.assertEqual(config["""hidden_size"""] , 768 ) def _a ( self : Dict ) -> Optional[int]: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = Path(_lowerCAmelCase ) / """a.txt""" filename.touch() self.assertEqual(get_file_from_repo(_lowerCAmelCase , """a.txt""" ) , str(_lowerCAmelCase ) ) self.assertIsNone(get_file_from_repo(_lowerCAmelCase , """b.txt""" ) )
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
1
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
1
import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def snake_case ( lowerCamelCase , lowerCamelCase=7 ): '''simple docstring''' __lowercase = None if token is not None: __lowercase = {"""Accept""": """application/vnd.github+json""", """Authorization""": F'Bearer {token}'} # The id of a workflow (not of a workflow run) __lowercase = """636036""" __lowercase = F'https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs' # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F'?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}' __lowercase = requests.get(lowerCamelCase , headers=lowerCamelCase ).json() return result["workflow_runs"] def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = get_daily_ci_runs(lowerCamelCase ) __lowercase = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": __lowercase = workflow_run["""id"""] break return workflow_run_id def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = get_last_daily_ci_runs(lowerCamelCase ) if workflow_run_id is not None: __lowercase = get_artifacts_links(worflow_run_id=lowerCamelCase , token=lowerCamelCase ) for artifact_name in artifact_names: if artifact_name in artifacts_links: __lowercase = artifacts_links[artifact_name] download_artifact( artifact_name=lowerCamelCase , artifact_url=lowerCamelCase , output_dir=lowerCamelCase , token=lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' get_last_daily_ci_artifacts(lowerCamelCase , lowerCamelCase , lowerCamelCase ) __lowercase = {} for artifact_name in artifact_names: __lowercase = os.path.join(lowerCamelCase , F'{artifact_name}.zip' ) if os.path.isfile(lowerCamelCase ): __lowercase = {} with zipfile.ZipFile(lowerCamelCase ) as z: for filename in z.namelist(): if not os.path.isdir(lowerCamelCase ): # read the file with z.open(lowerCamelCase ) as f: __lowercase = f.read().decode("""UTF-8""" ) return results
80
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
1
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __UpperCamelCase : Tuple = abspath(join(dirname(__file__), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def snake_case ( lowerCamelCase ): '''simple docstring''' config.addinivalue_line( """markers""" , """is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested""" ) config.addinivalue_line( """markers""" , """is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested""" ) config.addinivalue_line("""markers""" , """is_pipeline_test: mark test to run only when pipelines are tested""" ) config.addinivalue_line("""markers""" , """is_staging_test: mark test to run only in the staging environment""" ) config.addinivalue_line("""markers""" , """accelerate_tests: mark test that require accelerate""" ) config.addinivalue_line("""markers""" , """tool_tests: mark the tool tests that are run on their specific schedule""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowerCamelCase ) def snake_case ( lowerCamelCase ): '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main __lowercase = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(lowerCamelCase , id=lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if exitstatus == 5: __lowercase = 0 # Doctest custom flag to ignore output. __UpperCamelCase : Optional[Any] = doctest.register_optionflag("""IGNORE_RESULT""") __UpperCamelCase : List[Any] = doctest.OutputChecker class __UpperCamelCase ( _lowerCAmelCase ): def _a ( self : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict ) -> List[str]: """simple docstring""" if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : Optional[int] = CustomOutputChecker __UpperCamelCase : Optional[Any] = HfDoctestModule __UpperCamelCase : Any = HfDocTestParser
80
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def snake_case ( lowerCamelCase ): '''simple docstring''' return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = create_tensor(lowerCamelCase ) __lowercase = gather(lowerCamelCase ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = [state.process_index] __lowercase = gather_object(lowerCamelCase ) assert len(lowerCamelCase ) == state.num_processes, F'{gathered_obj}, {len(lowerCamelCase )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = create_tensor(lowerCamelCase ) __lowercase = broadcast(lowerCamelCase ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def snake_case ( lowerCamelCase ): '''simple docstring''' if state.is_main_process: __lowercase = torch.arange(state.num_processes + 1 ).to(state.device ) else: __lowercase = torch.arange(state.num_processes ).to(state.device ) __lowercase = pad_across_processes(lowerCamelCase ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def snake_case ( lowerCamelCase ): '''simple docstring''' if state.num_processes != 2: return __lowercase = create_tensor(lowerCamelCase ) __lowercase = reduce(lowerCamelCase , """sum""" ) __lowercase = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(lowerCamelCase , lowerCamelCase ), F'{reduced_tensor} != {truth_tensor}' def snake_case ( lowerCamelCase ): '''simple docstring''' if state.num_processes != 2: return __lowercase = create_tensor(lowerCamelCase ) __lowercase = reduce(lowerCamelCase , """mean""" ) __lowercase = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(lowerCamelCase , lowerCamelCase ), F'{reduced_tensor} != {truth_tensor}' def snake_case ( lowerCamelCase ): '''simple docstring''' main() def snake_case ( ): '''simple docstring''' __lowercase = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(lowerCamelCase ) state.print("""testing gather_object""" ) test_gather_object(lowerCamelCase ) state.print("""testing broadcast""" ) test_broadcast(lowerCamelCase ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(lowerCamelCase ) state.print("""testing reduce_sum""" ) test_reduce_sum(lowerCamelCase ) state.print("""testing reduce_mean""" ) test_reduce_mean(lowerCamelCase ) if __name__ == "__main__": main()
80
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class __UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[int] = ShapEPipeline __snake_case :Optional[int] = ['prompt'] __snake_case :str = ['prompt'] __snake_case :Tuple = [ 'num_images_per_prompt', 'num_inference_steps', 'generator', 'latents', 'guidance_scale', 'frame_size', 'output_type', 'return_dict', ] __snake_case :str = False @property def _a ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" return 32 @property def _a ( self : str ) -> str: """simple docstring""" return 32 @property def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" return self.time_input_dim * 4 @property def _a ( self : str ) -> Union[str, Any]: """simple docstring""" return 8 @property def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def _a ( self : int ) -> Optional[int]: """simple docstring""" torch.manual_seed(0 ) __lowercase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def _a ( self : Dict ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) __lowercase = { """num_attention_heads""": 2, """attention_head_dim""": 16, """embedding_dim""": self.time_input_dim, """num_embeddings""": 32, """embedding_proj_dim""": self.text_embedder_hidden_size, """time_embed_dim""": self.time_embed_dim, """num_layers""": 1, """clip_embed_dim""": self.time_input_dim * 2, """additional_embeddings""": 0, """time_embed_act_fn""": """gelu""", """norm_in_type""": """layer""", """encoder_hid_proj_type""": None, """added_emb_type""": None, } __lowercase = PriorTransformer(**_lowerCAmelCase ) return model @property def _a ( self : Any ) -> int: """simple docstring""" torch.manual_seed(0 ) __lowercase = { """param_shapes""": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), """d_latent""": self.time_input_dim, """d_hidden""": self.renderer_dim, """n_output""": 12, """background""": ( 0.1, 0.1, 0.1, ), } __lowercase = ShapERenderer(**_lowerCAmelCase ) return model def _a ( self : List[Any] ) -> str: """simple docstring""" __lowercase = self.dummy_prior __lowercase = self.dummy_text_encoder __lowercase = self.dummy_tokenizer __lowercase = self.dummy_renderer __lowercase = HeunDiscreteScheduler( beta_schedule="""exp""" , num_train_timesteps=1024 , prediction_type="""sample""" , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) __lowercase = { """prior""": prior, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """renderer""": renderer, """scheduler""": scheduler, } return components def _a ( self : Any , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any]=0 ) -> Dict: """simple docstring""" if str(_lowerCAmelCase ).startswith("""mps""" ): __lowercase = torch.manual_seed(_lowerCAmelCase ) else: __lowercase = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) __lowercase = { """prompt""": """horse""", """generator""": generator, """num_inference_steps""": 1, """frame_size""": 32, """output_type""": """np""", } return inputs def _a ( self : str ) -> Any: """simple docstring""" __lowercase = """cpu""" __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**_lowerCAmelCase ) __lowercase = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) __lowercase = output.images[0] __lowercase = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __lowercase = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _a ( self : Union[str, Any] ) -> int: """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _a ( self : Dict ) -> int: """simple docstring""" __lowercase = torch_device == """cpu""" __lowercase = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def _a ( self : Any ) -> str: """simple docstring""" __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**_lowerCAmelCase ) __lowercase = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = 1 __lowercase = 2 __lowercase = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: __lowercase = batch_size * [inputs[key]] __lowercase = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Dict ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/shap_e/test_shap_e_np_out.npy""" ) __lowercase = ShapEPipeline.from_pretrained("""openai/shap-e""" ) __lowercase = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) __lowercase = pipe( """a shark""" , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="""np""" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
80
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' if number < 0: raise ValueError("""number must not be negative""" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
80
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
1
import numpy as np class __UpperCamelCase : def __init__( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = (0, 0) __lowercase = None __lowercase = 0 __lowercase = 0 __lowercase = 0 def __eq__( self : List[Any] , _lowerCAmelCase : Dict ) -> Optional[Any]: """simple docstring""" return self.position == cell.position def _a ( self : Any ) -> Dict: """simple docstring""" print(self.position ) class __UpperCamelCase : def __init__( self : List[str] , _lowerCAmelCase : Any=(5, 5) ) -> Union[str, Any]: """simple docstring""" __lowercase = np.zeros(_lowerCAmelCase ) __lowercase = world_size[0] __lowercase = world_size[1] def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" print(self.w ) def _a ( self : Tuple , _lowerCAmelCase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] __lowercase = cell.position[0] __lowercase = cell.position[1] __lowercase = [] for n in neughbour_cord: __lowercase = current_x + n[0] __lowercase = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: __lowercase = Cell() __lowercase = (x, y) __lowercase = cell neighbours.append(_lowerCAmelCase ) return neighbours def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = [] _open.append(lowerCamelCase ) while _open: __lowercase = np.argmin([n.f for n in _open] ) __lowercase = _open[min_f] _closed.append(_open.pop(lowerCamelCase ) ) if current == goal: break for n in world.get_neigbours(lowerCamelCase ): for c in _closed: if c == n: continue __lowercase = current.g + 1 __lowercase , __lowercase = n.position __lowercase , __lowercase = goal.position __lowercase = (ya - ya) ** 2 + (xa - xa) ** 2 __lowercase = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(lowerCamelCase ) __lowercase = [] while current.parent is not None: path.append(current.position ) __lowercase = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __UpperCamelCase : Any = Gridworld() # Start position and goal __UpperCamelCase : Any = Cell() __UpperCamelCase : List[str] = (0, 0) __UpperCamelCase : Optional[int] = Cell() __UpperCamelCase : str = (4, 4) print(F'''path from {start.position} to {goal.position}''') __UpperCamelCase : str = astar(world, start, goal) # Just for visual reasons. for i in s: __UpperCamelCase : Any = 1 print(world.w)
80
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = len(lowerCamelCase ) while cur > 1: # Find the maximum number in arr __lowercase = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __lowercase = arr[mi::-1] + arr[mi + 1 : len(lowerCamelCase )] # Reverse whole list __lowercase = arr[cur - 1 :: -1] + arr[cur : len(lowerCamelCase )] cur -= 1 return arr if __name__ == "__main__": __UpperCamelCase : int = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
80
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
1
import os import sys import unittest __UpperCamelCase : Optional[int] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) __UpperCamelCase : Optional[int] = os.path.join("""tests""", """models""", """bert""", """test_modeling_bert.py""") __UpperCamelCase : str = os.path.join("""tests""", """models""", """blip""", """test_modeling_blip.py""") class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = get_test_to_tester_mapping(_lowerCAmelCase ) __lowercase = get_test_to_tester_mapping(_lowerCAmelCase ) __lowercase = {"""BertModelTest""": """BertModelTester"""} __lowercase = { """BlipModelTest""": """BlipModelTester""", """BlipTextImageModelTest""": """BlipTextImageModelsModelTester""", """BlipTextModelTest""": """BlipTextModelTester""", """BlipTextRetrievalModelTest""": """BlipTextRetrievalModelTester""", """BlipVQAModelTest""": """BlipVQAModelTester""", """BlipVisionModelTest""": """BlipVisionModelTester""", } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase = get_model_to_test_mapping(_lowerCAmelCase ) __lowercase = get_model_to_test_mapping(_lowerCAmelCase ) __lowercase = { """BertForMaskedLM""": ["""BertModelTest"""], """BertForMultipleChoice""": ["""BertModelTest"""], """BertForNextSentencePrediction""": ["""BertModelTest"""], """BertForPreTraining""": ["""BertModelTest"""], """BertForQuestionAnswering""": ["""BertModelTest"""], """BertForSequenceClassification""": ["""BertModelTest"""], """BertForTokenClassification""": ["""BertModelTest"""], """BertLMHeadModel""": ["""BertModelTest"""], """BertModel""": ["""BertModelTest"""], } __lowercase = { """BlipForConditionalGeneration""": ["""BlipTextImageModelTest"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTest"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTest"""], """BlipModel""": ["""BlipModelTest"""], """BlipTextModel""": ["""BlipTextModelTest"""], """BlipVisionModel""": ["""BlipVisionModelTest"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def _a ( self : List[str] ) -> int: """simple docstring""" __lowercase = get_model_to_tester_mapping(_lowerCAmelCase ) __lowercase = get_model_to_tester_mapping(_lowerCAmelCase ) __lowercase = { """BertForMaskedLM""": ["""BertModelTester"""], """BertForMultipleChoice""": ["""BertModelTester"""], """BertForNextSentencePrediction""": ["""BertModelTester"""], """BertForPreTraining""": ["""BertModelTester"""], """BertForQuestionAnswering""": ["""BertModelTester"""], """BertForSequenceClassification""": ["""BertModelTester"""], """BertForTokenClassification""": ["""BertModelTester"""], """BertLMHeadModel""": ["""BertModelTester"""], """BertModel""": ["""BertModelTester"""], } __lowercase = { """BlipForConditionalGeneration""": ["""BlipTextImageModelsModelTester"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTester"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTester"""], """BlipModel""": ["""BlipModelTester"""], """BlipTextModel""": ["""BlipTextModelTester"""], """BlipVisionModel""": ["""BlipVisionModelTester"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase )
80
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
1
from pathlib import Path import fire def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = Path(lowerCamelCase ) __lowercase = Path(lowerCamelCase ) dest_dir.mkdir(exist_ok=lowerCamelCase ) for path in src_dir.iterdir(): __lowercase = [x.rstrip() for x in list(path.open().readlines() )][:n] __lowercase = dest_dir.joinpath(path.name ) print(lowerCamelCase ) dest_path.open("""w""" ).write("""\n""".join(lowerCamelCase ) ) if __name__ == "__main__": fire.Fire(minify)
80
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
1
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness __UpperCamelCase : int = """\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } """ __UpperCamelCase : List[Any] = """\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper \"Evaluating Large Language Models Trained on Code\" (https://arxiv.org/abs/2107.03374). """ __UpperCamelCase : int = """ Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric(\"code_eval\") >>> test_cases = [\"assert add(2,3)==5\"] >>> candidates = [[\"def add(a,b): return a*b\", \"def add(a, b): return a+b\"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {'pass@1': 0.5, 'pass@2': 1.0} """ __UpperCamelCase : Union[str, Any] = """ ################################################################################ !!!WARNING!!! ################################################################################ The \"code_eval\" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper \"Evaluating Large Language Models Trained on Code\" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL=\"1\". Within Python you can to this with: >>> import os >>> os.environ[\"HF_ALLOW_CODE_EVAL\"] = \"1\" ################################################################################\ """ __UpperCamelCase : Any = """The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the \"Software\"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Value("""string""" ), } ) , homepage="""https://github.com/openai/human-eval""" , codebase_urls=["""https://github.com/openai/human-eval"""] , reference_urls=["""https://github.com/openai/human-eval"""] , license=_LICENSE , ) def _a ( self : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str]=[1, 10, 100] , _lowerCAmelCase : List[Any]=4 , _lowerCAmelCase : Optional[int]=3.0 ) -> List[Any]: """simple docstring""" if os.getenv("""HF_ALLOW_CODE_EVAL""" , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError("""This metric is currently not supported on Windows.""" ) with ThreadPoolExecutor(max_workers=_lowerCAmelCase ) as executor: __lowercase = [] __lowercase = Counter() __lowercase = 0 __lowercase = defaultdict(_lowerCAmelCase ) for task_id, (candidates, test_case) in enumerate(zip(_lowerCAmelCase , _lowerCAmelCase ) ): for candidate in candidates: __lowercase = candidate + """\n""" + test_case __lowercase = (test_program, timeout, task_id, completion_id[task_id]) __lowercase = executor.submit(_lowerCAmelCase , *_lowerCAmelCase ) futures.append(_lowerCAmelCase ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(_lowerCAmelCase ): __lowercase = future.result() results[result["task_id"]].append((result["""completion_id"""], result) ) __lowercase , __lowercase = [], [] for result in results.values(): result.sort() __lowercase = [r[1]["""passed"""] for r in result] total.append(len(_lowerCAmelCase ) ) correct.append(sum(_lowerCAmelCase ) ) __lowercase = np.array(_lowerCAmelCase ) __lowercase = np.array(_lowerCAmelCase ) __lowercase = k __lowercase = {F'pass@{k}': estimate_pass_at_k(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' def estimator(lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(lowerCamelCase , lowerCamelCase ): __lowercase = itertools.repeat(lowerCamelCase , len(lowerCamelCase ) ) else: assert len(lowerCamelCase ) == len(lowerCamelCase ) __lowercase = iter(lowerCamelCase ) return np.array([estimator(int(lowerCamelCase ) , int(lowerCamelCase ) , lowerCamelCase ) for n, c in zip(lowerCamelCase , lowerCamelCase )] )
80
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
1
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase = 10**-10 ): '''simple docstring''' __lowercase = a while True: __lowercase = Decimal(lowerCamelCase ) - ( Decimal(eval(lowerCamelCase ) ) / Decimal(eval(str(diff(lowerCamelCase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCamelCase ) ) < precision: # noqa: S307 return float(lowerCamelCase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'''The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}''') # Find root of polynomial print(F'''The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}''') # Find Square Root of 5 print(F'''The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}''') # Exponential Roots print(F'''The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}''')
80
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
1
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : Union[str, Any] = r""" Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. """ class __UpperCamelCase ( _lowerCAmelCase ): @add_start_docstrings(_lowerCAmelCase ) def __call__( self : int , _lowerCAmelCase : torch.LongTensor , _lowerCAmelCase : torch.FloatTensor , **_lowerCAmelCase : Optional[Any] ) -> bool: """simple docstring""" raise NotImplementedError("""StoppingCriteria needs to be subclassed""" ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] = None ) -> Tuple: """simple docstring""" __lowercase = max_length __lowercase = max_position_embeddings @add_start_docstrings(_lowerCAmelCase ) def __call__( self : List[str] , _lowerCAmelCase : torch.LongTensor , _lowerCAmelCase : torch.FloatTensor , **_lowerCAmelCase : int ) -> bool: """simple docstring""" __lowercase = input_ids.shape[-1] __lowercase = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( """This is a friendly reminder - the current text generation call will exceed the model's predefined """ F'maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ' """exceptions, performance degradation, or nothing at all.""" ) return is_done class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> Dict: """simple docstring""" warnings.warn( """The class `MaxNewTokensCriteria` is deprecated. """ F'Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ' """with `max_length = start_length + max_new_tokens` instead.""" , _lowerCAmelCase , ) __lowercase = start_length __lowercase = max_new_tokens __lowercase = start_length + max_new_tokens @add_start_docstrings(_lowerCAmelCase ) def __call__( self : List[Any] , _lowerCAmelCase : torch.LongTensor , _lowerCAmelCase : torch.FloatTensor , **_lowerCAmelCase : str ) -> bool: """simple docstring""" return input_ids.shape[-1] >= self.max_length class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : List[str] , _lowerCAmelCase : float , _lowerCAmelCase : Optional[float] = None ) -> Optional[int]: """simple docstring""" __lowercase = max_time __lowercase = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(_lowerCAmelCase ) def __call__( self : Tuple , _lowerCAmelCase : torch.LongTensor , _lowerCAmelCase : torch.FloatTensor , **_lowerCAmelCase : Any ) -> bool: """simple docstring""" return time.time() - self.initial_timestamp > self.max_time class __UpperCamelCase ( _lowerCAmelCase ): @add_start_docstrings(_lowerCAmelCase ) def __call__( self : Tuple , _lowerCAmelCase : torch.LongTensor , _lowerCAmelCase : torch.FloatTensor , **_lowerCAmelCase : str ) -> bool: """simple docstring""" return any(criteria(_lowerCAmelCase , _lowerCAmelCase ) for criteria in self ) @property def _a ( self : int ) -> Optional[int]: """simple docstring""" for stopping_criterium in self: if isinstance(_lowerCAmelCase , _lowerCAmelCase ): return stopping_criterium.max_length elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): return stopping_criterium.max_length return None def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = stopping_criteria.max_length __lowercase = deepcopy(lowerCamelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn("""You set different `max_length` for stopping criteria and `max_length` parameter""" , lowerCamelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=lowerCamelCase ) ) return new_stopping_criteria
80
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase : List[Any] = { """configuration_longformer""": [ """LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongformerConfig""", """LongformerOnnxConfig""", ], """tokenization_longformer""": ["""LongformerTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Any = ["""LongformerTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Optional[Any] = [ """LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongformerForMaskedLM""", """LongformerForMultipleChoice""", """LongformerForQuestionAnswering""", """LongformerForSequenceClassification""", """LongformerForTokenClassification""", """LongformerModel""", """LongformerPreTrainedModel""", """LongformerSelfAttention""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFLongformerForMaskedLM""", """TFLongformerForMultipleChoice""", """TFLongformerForQuestionAnswering""", """TFLongformerForSequenceClassification""", """TFLongformerForTokenClassification""", """TFLongformerModel""", """TFLongformerPreTrainedModel""", """TFLongformerSelfAttention""", ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys __UpperCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
1
__UpperCamelCase : List[str] = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {"""*""": op.mul, """/""": op.truediv, """+""": op.add, """-""": op.sub} __lowercase = Stack() __lowercase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(lowerCamelCase ) ) elif i in operators: # RULE 2 operator_stack.push(lowerCamelCase ) elif i == ")": # RULE 4 __lowercase = operator_stack.peek() operator_stack.pop() __lowercase = operand_stack.peek() operand_stack.pop() __lowercase = operand_stack.peek() operand_stack.pop() __lowercase = operators[opr](lowerCamelCase , lowerCamelCase ) operand_stack.push(lowerCamelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __UpperCamelCase : Union[str, Any] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
80
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __UpperCamelCase : Any = logging.get_logger(__name__) __UpperCamelCase : Any = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 'convnextv2' def __init__( self : List[Any] , _lowerCAmelCase : List[str]=3 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Dict=None , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : Any=1e-12 , _lowerCAmelCase : Union[str, Any]=0.0 , _lowerCAmelCase : str=224 , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : Any=None , **_lowerCAmelCase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(**_lowerCAmelCase ) __lowercase = num_channels __lowercase = patch_size __lowercase = num_stages __lowercase = [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes __lowercase = [3, 3, 9, 3] if depths is None else depths __lowercase = hidden_act __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = drop_path_rate __lowercase = image_size __lowercase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] __lowercase , __lowercase = get_aligned_output_features_output_indices( out_features=_lowerCAmelCase , out_indices=_lowerCAmelCase , stage_names=self.stage_names )
80
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : Optional[int] = TypeVar("""DatasetType""", Dataset, IterableDataset) def snake_case ( lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = "first_exhausted" , ): '''simple docstring''' from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError("""Unable to interleave an empty list of datasets.""" ) for i, dataset in enumerate(lowerCamelCase ): if not isinstance(lowerCamelCase , (Dataset, IterableDataset) ): if isinstance(lowerCamelCase , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ' """is an empty dataset dictionary.""" ) raise ValueError( F'Dataset at position {i} has at least one split: {list(lowerCamelCase )}\n' F'Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(lowerCamelCase ) )}\']' ) raise ValueError( F'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowerCamelCase ).__name__}.' ) if i == 0: __lowercase , __lowercase = ( (Dataset, IterableDataset) if isinstance(lowerCamelCase , lowerCamelCase ) else (IterableDataset, Dataset) ) elif not isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError( F'Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.' ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F'{stopping_strategy} is not supported. Please enter a valid stopping_strategy.' ) if dataset_type is Dataset: return _interleave_map_style_datasets( lowerCamelCase , lowerCamelCase , lowerCamelCase , info=lowerCamelCase , split=lowerCamelCase , stopping_strategy=lowerCamelCase ) else: return _interleave_iterable_datasets( lowerCamelCase , lowerCamelCase , lowerCamelCase , info=lowerCamelCase , split=lowerCamelCase , stopping_strategy=lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = 0 , ): '''simple docstring''' if not dsets: raise ValueError("""Unable to concatenate an empty list of datasets.""" ) for i, dataset in enumerate(lowerCamelCase ): if not isinstance(lowerCamelCase , (Dataset, IterableDataset) ): if isinstance(lowerCamelCase , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ' """is an empty dataset dictionary.""" ) raise ValueError( F'Dataset at position {i} has at least one split: {list(lowerCamelCase )}\n' F'Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(lowerCamelCase ) )}\']' ) raise ValueError( F'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowerCamelCase ).__name__}.' ) if i == 0: __lowercase , __lowercase = ( (Dataset, IterableDataset) if isinstance(lowerCamelCase , lowerCamelCase ) else (IterableDataset, Dataset) ) elif not isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError( F'Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.' ) if dataset_type is Dataset: return _concatenate_map_style_datasets(lowerCamelCase , info=lowerCamelCase , split=lowerCamelCase , axis=lowerCamelCase ) else: return _concatenate_iterable_datasets(lowerCamelCase , info=lowerCamelCase , split=lowerCamelCase , axis=lowerCamelCase )
80
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
1
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig __UpperCamelCase : Dict = logging.get_logger(__name__) # General docstring __UpperCamelCase : Any = """ResNetConfig""" # Base docstring __UpperCamelCase : Union[str, Any] = """microsoft/resnet-50""" __UpperCamelCase : List[Any] = [1, 2048, 7, 7] # Image classification docstring __UpperCamelCase : str = """microsoft/resnet-50""" __UpperCamelCase : List[str] = """tiger cat""" __UpperCamelCase : Union[str, Any] = [ """microsoft/resnet-50""", # See all resnet models at https://huggingface.co/models?filter=resnet ] class __UpperCamelCase ( nn.Module ): def __init__( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 3 , _lowerCAmelCase : int = 1 , _lowerCAmelCase : str = "relu" ) -> str: """simple docstring""" super().__init__() __lowercase = nn.Convad( _lowerCAmelCase , _lowerCAmelCase , kernel_size=_lowerCAmelCase , stride=_lowerCAmelCase , padding=kernel_size // 2 , bias=_lowerCAmelCase ) __lowercase = nn.BatchNormad(_lowerCAmelCase ) __lowercase = ACTaFN[activation] if activation is not None else nn.Identity() def _a ( self : List[Any] , _lowerCAmelCase : Tensor ) -> Tensor: """simple docstring""" __lowercase = self.convolution(_lowerCAmelCase ) __lowercase = self.normalization(_lowerCAmelCase ) __lowercase = self.activation(_lowerCAmelCase ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self : List[str] , _lowerCAmelCase : ResNetConfig ) -> str: """simple docstring""" super().__init__() __lowercase = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) __lowercase = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) __lowercase = config.num_channels def _a ( self : Tuple , _lowerCAmelCase : Tensor ) -> Tensor: """simple docstring""" __lowercase = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( """Make sure that the channel dimension of the pixel values match with the one set in the configuration.""" ) __lowercase = self.embedder(_lowerCAmelCase ) __lowercase = self.pooler(_lowerCAmelCase ) return embedding class __UpperCamelCase ( nn.Module ): def __init__( self : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 2 ) -> Tuple: """simple docstring""" super().__init__() __lowercase = nn.Convad(_lowerCAmelCase , _lowerCAmelCase , kernel_size=1 , stride=_lowerCAmelCase , bias=_lowerCAmelCase ) __lowercase = nn.BatchNormad(_lowerCAmelCase ) def _a ( self : int , _lowerCAmelCase : Tensor ) -> Tensor: """simple docstring""" __lowercase = self.convolution(_lowerCAmelCase ) __lowercase = self.normalization(_lowerCAmelCase ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self : str , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 1 , _lowerCAmelCase : str = "relu" ) -> Optional[int]: """simple docstring""" super().__init__() __lowercase = in_channels != out_channels or stride != 1 __lowercase = ( ResNetShortCut(_lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase ) if should_apply_shortcut else nn.Identity() ) __lowercase = nn.Sequential( ResNetConvLayer(_lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase ) , ResNetConvLayer(_lowerCAmelCase , _lowerCAmelCase , activation=_lowerCAmelCase ) , ) __lowercase = ACTaFN[activation] def _a ( self : List[str] , _lowerCAmelCase : Dict ) -> Tuple: """simple docstring""" __lowercase = hidden_state __lowercase = self.layer(_lowerCAmelCase ) __lowercase = self.shortcut(_lowerCAmelCase ) hidden_state += residual __lowercase = self.activation(_lowerCAmelCase ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 1 , _lowerCAmelCase : str = "relu" , _lowerCAmelCase : int = 4 ) -> Any: """simple docstring""" super().__init__() __lowercase = in_channels != out_channels or stride != 1 __lowercase = out_channels // reduction __lowercase = ( ResNetShortCut(_lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase ) if should_apply_shortcut else nn.Identity() ) __lowercase = nn.Sequential( ResNetConvLayer(_lowerCAmelCase , _lowerCAmelCase , kernel_size=1 ) , ResNetConvLayer(_lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase ) , ResNetConvLayer(_lowerCAmelCase , _lowerCAmelCase , kernel_size=1 , activation=_lowerCAmelCase ) , ) __lowercase = ACTaFN[activation] def _a ( self : Optional[int] , _lowerCAmelCase : List[str] ) -> Tuple: """simple docstring""" __lowercase = hidden_state __lowercase = self.layer(_lowerCAmelCase ) __lowercase = self.shortcut(_lowerCAmelCase ) hidden_state += residual __lowercase = self.activation(_lowerCAmelCase ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self : Optional[int] , _lowerCAmelCase : ResNetConfig , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 2 , _lowerCAmelCase : int = 2 , ) -> List[str]: """simple docstring""" super().__init__() __lowercase = ResNetBottleNeckLayer if config.layer_type == """bottleneck""" else ResNetBasicLayer __lowercase = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(_lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase , activation=config.hidden_act ) , *[layer(_lowerCAmelCase , _lowerCAmelCase , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def _a ( self : Optional[int] , _lowerCAmelCase : Tensor ) -> Tensor: """simple docstring""" __lowercase = input for layer in self.layers: __lowercase = layer(_lowerCAmelCase ) return hidden_state class __UpperCamelCase ( nn.Module ): def __init__( self : List[str] , _lowerCAmelCase : ResNetConfig ) -> List[str]: """simple docstring""" super().__init__() __lowercase = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( _lowerCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) __lowercase = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_lowerCAmelCase , config.depths[1:] ): self.stages.append(ResNetStage(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , depth=_lowerCAmelCase ) ) def _a ( self : Any , _lowerCAmelCase : Tensor , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = True ) -> BaseModelOutputWithNoAttention: """simple docstring""" __lowercase = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __lowercase = hidden_states + (hidden_state,) __lowercase = stage_module(_lowerCAmelCase ) if output_hidden_states: __lowercase = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=_lowerCAmelCase , hidden_states=_lowerCAmelCase , ) class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Optional[Any] = ResNetConfig __snake_case :str = 'resnet' __snake_case :Dict = 'pixel_values' __snake_case :Any = True def _a ( self : List[str] , _lowerCAmelCase : Any ) -> Optional[int]: """simple docstring""" if isinstance(_lowerCAmelCase , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="""fan_out""" , nonlinearity="""relu""" ) elif isinstance(_lowerCAmelCase , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _a ( self : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=False ) -> int: """simple docstring""" if isinstance(_lowerCAmelCase , _lowerCAmelCase ): __lowercase = value __UpperCamelCase : str = r""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`ResNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ __UpperCamelCase : List[str] = r""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( 'The bare ResNet model outputting raw features without any specific head on top.' , _lowerCAmelCase , ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : List[Any] , _lowerCAmelCase : Union[str, Any] ) -> str: """simple docstring""" super().__init__(_lowerCAmelCase ) __lowercase = config __lowercase = ResNetEmbeddings(_lowerCAmelCase ) __lowercase = ResNetEncoder(_lowerCAmelCase ) __lowercase = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_lowerCAmelCase , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _a ( self : Dict , _lowerCAmelCase : Tensor , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[bool] = None ) -> BaseModelOutputWithPoolingAndNoAttention: """simple docstring""" __lowercase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowercase = return_dict if return_dict is not None else self.config.use_return_dict __lowercase = self.embedder(_lowerCAmelCase ) __lowercase = self.encoder( _lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase ) __lowercase = encoder_outputs[0] __lowercase = self.pooler(_lowerCAmelCase ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_lowerCAmelCase , pooler_output=_lowerCAmelCase , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ' , _lowerCAmelCase , ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Dict , _lowerCAmelCase : int ) -> Optional[int]: """simple docstring""" super().__init__(_lowerCAmelCase ) __lowercase = config.num_labels __lowercase = ResNetModel(_lowerCAmelCase ) # classification head __lowercase = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_lowerCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _a ( self : int , _lowerCAmelCase : Optional[torch.FloatTensor] = None , _lowerCAmelCase : Optional[torch.LongTensor] = None , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: """simple docstring""" __lowercase = return_dict if return_dict is not None else self.config.use_return_dict __lowercase = self.resnet(_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase ) __lowercase = outputs.pooler_output if return_dict else outputs[1] __lowercase = self.classifier(_lowerCAmelCase ) __lowercase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __lowercase = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __lowercase = """single_label_classification""" else: __lowercase = """multi_label_classification""" if self.config.problem_type == "regression": __lowercase = MSELoss() if self.num_labels == 1: __lowercase = loss_fct(logits.squeeze() , labels.squeeze() ) else: __lowercase = loss_fct(_lowerCAmelCase , _lowerCAmelCase ) elif self.config.problem_type == "single_label_classification": __lowercase = CrossEntropyLoss() __lowercase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __lowercase = BCEWithLogitsLoss() __lowercase = loss_fct(_lowerCAmelCase , _lowerCAmelCase ) if not return_dict: __lowercase = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_lowerCAmelCase , logits=_lowerCAmelCase , hidden_states=outputs.hidden_states ) @add_start_docstrings( '\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n ' , _lowerCAmelCase , ) class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): def __init__( self : Tuple , _lowerCAmelCase : Tuple ) -> str: """simple docstring""" super().__init__(_lowerCAmelCase ) super()._init_backbone(_lowerCAmelCase ) __lowercase = [config.embedding_size] + config.hidden_sizes __lowercase = ResNetEmbeddings(_lowerCAmelCase ) __lowercase = ResNetEncoder(_lowerCAmelCase ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_lowerCAmelCase ) @replace_return_docstrings(output_type=_lowerCAmelCase , config_class=_CONFIG_FOR_DOC ) def _a ( self : int , _lowerCAmelCase : Tensor , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[bool] = None ) -> BackboneOutput: """simple docstring""" __lowercase = return_dict if return_dict is not None else self.config.use_return_dict __lowercase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowercase = self.embedder(_lowerCAmelCase ) __lowercase = self.encoder(_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase ) __lowercase = outputs.hidden_states __lowercase = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: __lowercase = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=_lowerCAmelCase , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=_lowerCAmelCase , )
80
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
1
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = s.rsplit(lowerCamelCase , lowerCamelCase ) return new.join(lowerCamelCase ) def snake_case ( lowerCamelCase ): '''simple docstring''' return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} __lowercase = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: __lowercase = key.replace(F'{group_key}.' , F'{group_key}.group.' ) if "res_path" in key: __lowercase = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): __lowercase = rreplace(lowerCamelCase , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): __lowercase = rreplace(lowerCamelCase , """.b""" , """.bias""" , 1 ) __lowercase = value.float() return upgrade @torch.no_grad() def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=True ): '''simple docstring''' from dall_e import Encoder __lowercase = Encoder() if os.path.exists(lowerCamelCase ): __lowercase = torch.load(lowerCamelCase ) else: __lowercase = torch.hub.load_state_dict_from_url(lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ): __lowercase = ckpt.state_dict() encoder.load_state_dict(lowerCamelCase ) if config_path is not None: __lowercase = FlavaImageCodebookConfig.from_pretrained(lowerCamelCase ) else: __lowercase = FlavaImageCodebookConfig() __lowercase = FlavaImageCodebook(lowerCamelCase ).eval() __lowercase = encoder.state_dict() __lowercase = upgrade_state_dict(lowerCamelCase ) hf_model.load_state_dict(lowerCamelCase ) __lowercase = hf_model.state_dict() __lowercase = count_parameters(lowerCamelCase ) __lowercase = count_parameters(lowerCamelCase ) assert torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) if save_checkpoint: hf_model.save_pretrained(lowerCamelCase ) else: return hf_state_dict if __name__ == "__main__": __UpperCamelCase : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") __UpperCamelCase : Optional[Any] = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
80
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
1
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
1
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel __UpperCamelCase : Dict = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } __UpperCamelCase : Optional[int] = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def snake_case ( lowerCamelCase , lowerCamelCase=False ): '''simple docstring''' __lowercase , __lowercase = create_model( """HTSAT-tiny""" , """roberta""" , lowerCamelCase , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=lowerCamelCase , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} __lowercase = r""".*sequential.(\d+).*""" __lowercase = r""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __lowercase = key.replace(lowerCamelCase , lowerCamelCase ) if re.match(lowerCamelCase , lowerCamelCase ): # replace sequential layers with list __lowercase = re.match(lowerCamelCase , lowerCamelCase ).group(1 ) __lowercase = key.replace(F'sequential.{sequential_layer}.' , F'layers.{int(lowerCamelCase )//3}.linear.' ) elif re.match(lowerCamelCase , lowerCamelCase ): __lowercase = int(re.match(lowerCamelCase , lowerCamelCase ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... __lowercase = 1 if projecton_layer == 0 else 2 __lowercase = key.replace(F'_projection.{projecton_layer}.' , F'_projection.linear{transformers_projection_layer}.' ) if "audio" and "qkv" in key: # split qkv into query key and value __lowercase = value __lowercase = mixed_qkv.size(0 ) // 3 __lowercase = mixed_qkv[:qkv_dim] __lowercase = mixed_qkv[qkv_dim : qkv_dim * 2] __lowercase = mixed_qkv[qkv_dim * 2 :] __lowercase = query_layer __lowercase = key_layer __lowercase = value_layer else: __lowercase = value return model_state_dict def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): '''simple docstring''' __lowercase , __lowercase = init_clap(lowerCamelCase , enable_fusion=lowerCamelCase ) clap_model.eval() __lowercase = clap_model.state_dict() __lowercase = rename_state_dict(lowerCamelCase ) __lowercase = ClapConfig() __lowercase = enable_fusion __lowercase = ClapModel(lowerCamelCase ) # ignore the spectrogram embedding layer model.load_state_dict(lowerCamelCase , strict=lowerCamelCase ) model.save_pretrained(lowerCamelCase ) transformers_config.save_pretrained(lowerCamelCase ) if __name__ == "__main__": __UpperCamelCase : List[str] = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") __UpperCamelCase : Optional[Any] = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : Union[str, Any] = { """bert-base-uncased""": """https://huggingface.co/bert-base-uncased/resolve/main/config.json""", """bert-large-uncased""": """https://huggingface.co/bert-large-uncased/resolve/main/config.json""", """bert-base-cased""": """https://huggingface.co/bert-base-cased/resolve/main/config.json""", """bert-large-cased""": """https://huggingface.co/bert-large-cased/resolve/main/config.json""", """bert-base-multilingual-uncased""": """https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json""", """bert-base-multilingual-cased""": """https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json""", """bert-base-chinese""": """https://huggingface.co/bert-base-chinese/resolve/main/config.json""", """bert-base-german-cased""": """https://huggingface.co/bert-base-german-cased/resolve/main/config.json""", """bert-large-uncased-whole-word-masking""": ( """https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json""" ), """bert-large-cased-whole-word-masking""": ( """https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json""" ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( """https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json""" ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( """https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json""" ), """bert-base-cased-finetuned-mrpc""": """https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json""", """bert-base-german-dbmdz-cased""": """https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json""", """bert-base-german-dbmdz-uncased""": """https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json""", """cl-tohoku/bert-base-japanese""": """https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json""", """cl-tohoku/bert-base-japanese-whole-word-masking""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json""" ), """cl-tohoku/bert-base-japanese-char""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json""" ), """cl-tohoku/bert-base-japanese-char-whole-word-masking""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json""" ), """TurkuNLP/bert-base-finnish-cased-v1""": ( """https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json""" ), """TurkuNLP/bert-base-finnish-uncased-v1""": ( """https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json""" ), """wietsedv/bert-base-dutch-cased""": """https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json""", # See all BERT models at https://huggingface.co/models?filter=bert } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :int = 'bert' def __init__( self : Union[str, Any] , _lowerCAmelCase : Union[str, Any]=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : List[Any]=12 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : List[str]=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : Optional[Any]=512 , _lowerCAmelCase : Dict=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : Dict=1e-12 , _lowerCAmelCase : Optional[int]=0 , _lowerCAmelCase : Optional[Any]="absolute" , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=None , **_lowerCAmelCase : Optional[Any] , ) -> Optional[int]: """simple docstring""" super().__init__(pad_token_id=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = classifier_dropout class __UpperCamelCase ( _lowerCAmelCase ): @property def _a ( self : Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": __lowercase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowercase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
1
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __UpperCamelCase : Optional[Any] = logging.get_logger(__name__) __UpperCamelCase : int = { """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Any = 'deta' __snake_case :Union[str, Any] = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Optional[Any] , _lowerCAmelCase : Dict=None , _lowerCAmelCase : Optional[Any]=900 , _lowerCAmelCase : Optional[Any]=2048 , _lowerCAmelCase : int=6 , _lowerCAmelCase : List[str]=2048 , _lowerCAmelCase : List[Any]=8 , _lowerCAmelCase : str=6 , _lowerCAmelCase : Dict=1024 , _lowerCAmelCase : str=8 , _lowerCAmelCase : str=0.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[Any]="relu" , _lowerCAmelCase : Optional[Any]=256 , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : Dict=0.0 , _lowerCAmelCase : Dict=0.0 , _lowerCAmelCase : Dict=0.02 , _lowerCAmelCase : Optional[int]=1.0 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : str=False , _lowerCAmelCase : Optional[int]="sine" , _lowerCAmelCase : int=5 , _lowerCAmelCase : int=4 , _lowerCAmelCase : List[str]=4 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=300 , _lowerCAmelCase : str=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Any=1 , _lowerCAmelCase : Dict=5 , _lowerCAmelCase : Dict=2 , _lowerCAmelCase : Any=1 , _lowerCAmelCase : str=1 , _lowerCAmelCase : Union[str, Any]=5 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Optional[Any]=0.25 , **_lowerCAmelCase : Any , ) -> Tuple: """simple docstring""" if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) __lowercase = CONFIG_MAPPING["""resnet"""](out_features=["""stage2""", """stage3""", """stage4"""] ) else: if isinstance(_lowerCAmelCase , _lowerCAmelCase ): __lowercase = backbone_config.pop("""model_type""" ) __lowercase = CONFIG_MAPPING[backbone_model_type] __lowercase = config_class.from_dict(_lowerCAmelCase ) __lowercase = backbone_config __lowercase = num_queries __lowercase = max_position_embeddings __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = init_xavier_std __lowercase = encoder_layerdrop __lowercase = auxiliary_loss __lowercase = position_embedding_type # deformable attributes __lowercase = num_feature_levels __lowercase = encoder_n_points __lowercase = decoder_n_points __lowercase = two_stage __lowercase = two_stage_num_proposals __lowercase = with_box_refine __lowercase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("""If two_stage is True, with_box_refine must be True.""" ) # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = mask_loss_coefficient __lowercase = dice_loss_coefficient __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient __lowercase = focal_alpha super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def _a ( self : Optional[Any] ) -> int: """simple docstring""" return self.encoder_attention_heads @property def _a ( self : Dict ) -> int: """simple docstring""" return self.d_model def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = copy.deepcopy(self.__dict__ ) __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output
80
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
1
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor __UpperCamelCase : Tuple = random.Random() def snake_case ( lowerCamelCase , lowerCamelCase=1.0 , lowerCamelCase=None , lowerCamelCase=None ): '''simple docstring''' if rng is None: __lowercase = global_rng __lowercase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __UpperCamelCase ( unittest.TestCase ): def __init__( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Any=7 , _lowerCAmelCase : Dict=400 , _lowerCAmelCase : Any=2000 , _lowerCAmelCase : Dict=24 , _lowerCAmelCase : Optional[int]=24 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : int=1_6000 , _lowerCAmelCase : Any=True , _lowerCAmelCase : List[Any]=True , ) -> str: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = min_seq_length __lowercase = max_seq_length __lowercase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowercase = feature_size __lowercase = num_mel_bins __lowercase = padding_value __lowercase = sampling_rate __lowercase = return_attention_mask __lowercase = do_normalize def _a ( self : List[str] ) -> str: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _a ( self : int , _lowerCAmelCase : Tuple=False , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _flatten(_lowerCAmelCase : Optional[Any] ): return list(itertools.chain(*_lowerCAmelCase ) ) if equal_length: __lowercase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowercase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowercase = [np.asarray(_lowerCAmelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = SpeechaTextFeatureExtractor if is_speech_available() else None def _a ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase = SpeechaTextFeatureExtractionTester(self ) def _a ( self : Tuple , _lowerCAmelCase : List[Any] ) -> Tuple: """simple docstring""" self.assertTrue(np.all(np.mean(_lowerCAmelCase , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(_lowerCAmelCase , axis=0 ) - 1 ) < 1e-3 ) ) def _a ( self : int ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = [np.asarray(_lowerCAmelCase ) for speech_input in speech_inputs] # Test feature size __lowercase = feature_extractor(_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowercase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __lowercase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) ) # Test batched __lowercase = feature_extractor(_lowerCAmelCase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(_lowerCAmelCase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assertTrue(np.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowercase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowercase = np.asarray(_lowerCAmelCase ) __lowercase = feature_extractor(_lowerCAmelCase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(_lowerCAmelCase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assertTrue(np.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(_lowerCAmelCase , _lowerCAmelCase ): __lowercase = feature_extractor( _lowerCAmelCase , padding=_lowerCAmelCase , max_length=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(_lowerCAmelCase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(_lowerCAmelCase , _lowerCAmelCase ): __lowercase = feature_extractor( _lowerCAmelCase , max_length=_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors="""np""" , return_attention_mask=_lowerCAmelCase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(_lowerCAmelCase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = feature_extractor( _lowerCAmelCase , padding="""max_length""" , max_length=4 , truncation=_lowerCAmelCase , return_tensors="""np""" , return_attention_mask=_lowerCAmelCase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def _a ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = feature_extractor( _lowerCAmelCase , padding="""longest""" , max_length=4 , truncation=_lowerCAmelCase , return_tensors="""np""" , return_attention_mask=_lowerCAmelCase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowercase = feature_extractor( _lowerCAmelCase , padding="""longest""" , max_length=16 , truncation=_lowerCAmelCase , return_tensors="""np""" , return_attention_mask=_lowerCAmelCase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def _a ( self : Dict ) -> Union[str, Any]: """simple docstring""" import torch __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = np.random.rand(100 , 32 ).astype(np.floataa ) __lowercase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _a ( self : List[str] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" from datasets import load_dataset __lowercase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowercase = ds.sort("""id""" ).select(range(_lowerCAmelCase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _a ( self : int ) -> Any: """simple docstring""" __lowercase = np.array([ -1.5_745, -1.7_713, -1.7_020, -1.6_069, -1.2_250, -1.1_105, -0.9_072, -0.8_241, -1.2_310, -0.8_098, -0.3_320, -0.4_101, -0.7_985, -0.4_996, -0.8_213, -0.9_128, -1.0_420, -1.1_286, -1.0_440, -0.7_999, -0.8_405, -1.2_275, -1.5_443, -1.4_625, ] ) # fmt: on __lowercase = self._load_datasamples(1 ) __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = feature_extractor(_lowerCAmelCase , return_tensors="""pt""" ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , _lowerCAmelCase , atol=1e-4 ) )
80
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
1
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class __UpperCamelCase ( nn.Module ): __snake_case :int __snake_case :int __snake_case :float = 0.0 __snake_case :int = 1 __snake_case :int = 1 __snake_case :bool = True __snake_case :bool = False __snake_case :bool = False __snake_case :bool = False __snake_case :jnp.dtype = jnp.floataa def _a ( self : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = [] __lowercase = [] for i in range(self.num_layers ): __lowercase = self.in_channels if i == 0 else self.out_channels __lowercase = FlaxResnetBlockaD( in_channels=_lowerCAmelCase , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_lowerCAmelCase ) __lowercase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_lowerCAmelCase ) __lowercase = resnets __lowercase = attentions if self.add_downsample: __lowercase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=True ) -> Dict: """simple docstring""" __lowercase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowercase = resnet(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) __lowercase = attn(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) output_states += (hidden_states,) if self.add_downsample: __lowercase = self.downsamplers_a(_lowerCAmelCase ) output_states += (hidden_states,) return hidden_states, output_states class __UpperCamelCase ( nn.Module ): __snake_case :int __snake_case :int __snake_case :float = 0.0 __snake_case :int = 1 __snake_case :bool = True __snake_case :jnp.dtype = jnp.floataa def _a ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = [] for i in range(self.num_layers ): __lowercase = self.in_channels if i == 0 else self.out_channels __lowercase = FlaxResnetBlockaD( in_channels=_lowerCAmelCase , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_lowerCAmelCase ) __lowercase = resnets if self.add_downsample: __lowercase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self : Union[str, Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int]=True ) -> Union[str, Any]: """simple docstring""" __lowercase = () for resnet in self.resnets: __lowercase = resnet(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) output_states += (hidden_states,) if self.add_downsample: __lowercase = self.downsamplers_a(_lowerCAmelCase ) output_states += (hidden_states,) return hidden_states, output_states class __UpperCamelCase ( nn.Module ): __snake_case :int __snake_case :int __snake_case :int __snake_case :float = 0.0 __snake_case :int = 1 __snake_case :int = 1 __snake_case :bool = True __snake_case :bool = False __snake_case :bool = False __snake_case :bool = False __snake_case :jnp.dtype = jnp.floataa def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = [] __lowercase = [] for i in range(self.num_layers ): __lowercase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowercase = self.prev_output_channel if i == 0 else self.out_channels __lowercase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_lowerCAmelCase ) __lowercase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_lowerCAmelCase ) __lowercase = resnets __lowercase = attentions if self.add_upsample: __lowercase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict=True ) -> str: """simple docstring""" for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowercase = res_hidden_states_tuple[-1] __lowercase = res_hidden_states_tuple[:-1] __lowercase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowercase = resnet(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) __lowercase = attn(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) if self.add_upsample: __lowercase = self.upsamplers_a(_lowerCAmelCase ) return hidden_states class __UpperCamelCase ( nn.Module ): __snake_case :int __snake_case :int __snake_case :int __snake_case :float = 0.0 __snake_case :int = 1 __snake_case :bool = True __snake_case :jnp.dtype = jnp.floataa def _a ( self : int ) -> Any: """simple docstring""" __lowercase = [] for i in range(self.num_layers ): __lowercase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowercase = self.prev_output_channel if i == 0 else self.out_channels __lowercase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_lowerCAmelCase ) __lowercase = resnets if self.add_upsample: __lowercase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str]=True ) -> Tuple: """simple docstring""" for resnet in self.resnets: # pop res hidden states __lowercase = res_hidden_states_tuple[-1] __lowercase = res_hidden_states_tuple[:-1] __lowercase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowercase = resnet(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) if self.add_upsample: __lowercase = self.upsamplers_a(_lowerCAmelCase ) return hidden_states class __UpperCamelCase ( nn.Module ): __snake_case :int __snake_case :float = 0.0 __snake_case :int = 1 __snake_case :int = 1 __snake_case :bool = False __snake_case :bool = False __snake_case :jnp.dtype = jnp.floataa def _a ( self : str ) -> List[Any]: """simple docstring""" __lowercase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowercase = [] for _ in range(self.num_layers ): __lowercase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(_lowerCAmelCase ) __lowercase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(_lowerCAmelCase ) __lowercase = resnets __lowercase = attentions def __call__( self : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any]=True ) -> List[str]: """simple docstring""" __lowercase = self.resnets[0](_lowerCAmelCase , _lowerCAmelCase ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowercase = attn(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) __lowercase = resnet(_lowerCAmelCase , _lowerCAmelCase , deterministic=_lowerCAmelCase ) return hidden_states
80
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
1
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_lowerCAmelCase ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , *_lowerCAmelCase : Any , **_lowerCAmelCase : int ) -> Dict: """simple docstring""" super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == """tf""" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : Any=None , _lowerCAmelCase : Dict=None ) -> int: """simple docstring""" __lowercase = {} __lowercase = {} if prompt is not None: __lowercase = prompt if generate_kwargs is not None: __lowercase = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: __lowercase = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( """'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter,""" """ please use only one""" ) __lowercase = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self : Union[str, Any] , _lowerCAmelCase : Union[str, List[str], "Image.Image", List["Image.Image"]] , **_lowerCAmelCase : Any ) -> Tuple: """simple docstring""" return super().__call__(_lowerCAmelCase , **_lowerCAmelCase ) def _a ( self : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any=None ) -> Any: """simple docstring""" __lowercase = load_image(_lowerCAmelCase ) if prompt is not None: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise ValueError( F'Received an invalid text input, got - {type(_lowerCAmelCase )} - but expected a single string. ' """Note also that one single text can be provided for conditional image to text generation.""" ) __lowercase = self.model.config.model_type if model_type == "git": __lowercase = self.image_processor(images=_lowerCAmelCase , return_tensors=self.framework ) __lowercase = self.tokenizer(text=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ).input_ids __lowercase = [self.tokenizer.cls_token_id] + input_ids __lowercase = torch.tensor(_lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"""input_ids""": input_ids} ) elif model_type == "pix2struct": __lowercase = self.image_processor(images=_lowerCAmelCase , header_text=_lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation __lowercase = self.image_processor(images=_lowerCAmelCase , return_tensors=self.framework ) __lowercase = self.tokenizer(_lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(_lowerCAmelCase ) else: raise ValueError(F'Model type {model_type} does not support conditional text generation' ) else: __lowercase = self.image_processor(images=_lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: __lowercase = None return model_inputs def _a ( self : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" if ( "input_ids" in model_inputs and isinstance(model_inputs["""input_ids"""] , _lowerCAmelCase ) and all(x is None for x in model_inputs["""input_ids"""] ) ): __lowercase = None if generate_kwargs is None: __lowercase = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. __lowercase = model_inputs.pop(self.model.main_input_name ) __lowercase = self.model.generate(_lowerCAmelCase , **_lowerCAmelCase , **_lowerCAmelCase ) return model_outputs def _a ( self : List[str] , _lowerCAmelCase : Union[str, Any] ) -> int: """simple docstring""" __lowercase = [] for output_ids in model_outputs: __lowercase = { """generated_text""": self.tokenizer.decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , ) } records.append(_lowerCAmelCase ) return records
80
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = tempfile.mkdtemp() # fmt: off __lowercase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] __lowercase = {"""unk_token""": """<unk>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowerCAmelCase ) ) __lowercase = { """do_resize""": True, """size""": 20, """do_center_crop""": True, """crop_size""": 18, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } __lowercase = os.path.join(self.tmpdirname , _lowerCAmelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] , **_lowerCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **_lowerCAmelCase ) def _a ( self : int , **_lowerCAmelCase : int ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **_lowerCAmelCase ) def _a ( self : str , **_lowerCAmelCase : List[str] ) -> List[str]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def _a ( self : str ) -> Tuple: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __lowercase = [Image.fromarray(np.moveaxis(_lowerCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = self.get_image_processor() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor_slow.save_pretrained(self.tmpdirname ) __lowercase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_lowerCAmelCase ) __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor_fast.save_pretrained(self.tmpdirname ) __lowercase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _lowerCAmelCase ) self.assertIsInstance(processor_fast.tokenizer , _lowerCAmelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _lowerCAmelCase ) self.assertIsInstance(processor_fast.image_processor , _lowerCAmelCase ) def _a ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __lowercase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) __lowercase = self.get_image_processor(do_normalize=_lowerCAmelCase ) __lowercase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=_lowerCAmelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _lowerCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) __lowercase = self.prepare_image_inputs() __lowercase = image_processor(_lowerCAmelCase , return_tensors="""np""" ) __lowercase = processor(images=_lowerCAmelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) __lowercase = """lower newer""" __lowercase = processor(text=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = tokenizer(_lowerCAmelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _a ( self : Tuple ) -> int: """simple docstring""" __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) __lowercase = """lower newer""" __lowercase = self.prepare_image_inputs() __lowercase = processor(text=_lowerCAmelCase , images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def _a ( self : str ) -> Any: """simple docstring""" __lowercase = """google/owlvit-base-patch32""" __lowercase = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) __lowercase = ["""cat""", """nasa badge"""] __lowercase = processor(text=_lowerCAmelCase ) __lowercase = 16 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = """google/owlvit-base-patch32""" __lowercase = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) __lowercase = [["""cat""", """nasa badge"""], ["""person"""]] __lowercase = processor(text=_lowerCAmelCase ) __lowercase = 16 __lowercase = len(_lowerCAmelCase ) __lowercase = max([len(_lowerCAmelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def _a ( self : Optional[Any] ) -> Any: """simple docstring""" __lowercase = """google/owlvit-base-patch32""" __lowercase = OwlViTProcessor.from_pretrained(_lowerCAmelCase ) __lowercase = ["""cat""", """nasa badge"""] __lowercase = processor(text=_lowerCAmelCase ) __lowercase = 16 __lowercase = inputs["""input_ids"""] __lowercase = [ [4_9406, 2368, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9406, 6841, 1_1301, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) __lowercase = self.prepare_image_inputs() __lowercase = self.prepare_image_inputs() __lowercase = processor(images=_lowerCAmelCase , query_images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = OwlViTProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) __lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowercase = processor.batch_decode(_lowerCAmelCase ) __lowercase = tokenizer.batch_decode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase )
80
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
1
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path __UpperCamelCase : Optional[Any] = Path(__file__).resolve().parents[3] / """src""" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) __UpperCamelCase : str = {"""base""": """patrickvonplaten/wav2vec2_tiny_random""", """robust""": """patrickvonplaten/wav2vec2_tiny_random_robust"""} __UpperCamelCase : List[str] = """zero2""" __UpperCamelCase : List[Any] = """zero3""" __UpperCamelCase : Optional[int] = [ZEROa, ZEROa] def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = parameterized.to_safe_name("""_""".join(str(lowerCamelCase ) for x in param.args ) ) return F'{func.__name__}_{param_based_name}' # Cartesian-product of zero stages with models to test __UpperCamelCase : Union[str, Any] = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class __UpperCamelCase ( _lowerCAmelCase ): @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def _a ( self : Dict , _lowerCAmelCase : int , _lowerCAmelCase : Dict ) -> Union[str, Any]: """simple docstring""" self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> Tuple: """simple docstring""" self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def _a ( self : Tuple , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCAmelCase , name_func=_lowerCAmelCase ) def _a ( self : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict ) -> str: """simple docstring""" self.run_and_check( stage=_lowerCAmelCase , model=_lowerCAmelCase , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[Any] ) -> Any: """simple docstring""" pass def _a ( self : str , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : int = 10 , _lowerCAmelCase : bool = True , _lowerCAmelCase : bool = True , _lowerCAmelCase : bool = True , ) -> str: """simple docstring""" __lowercase = models[model] __lowercase = self.run_trainer( stage=_lowerCAmelCase , model_name=_lowerCAmelCase , eval_steps=_lowerCAmelCase , num_train_epochs=1 , distributed=_lowerCAmelCase , fpaa=_lowerCAmelCase , ) self.do_checks(_lowerCAmelCase ) return output_dir def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : int = 10 , _lowerCAmelCase : int = 1 , _lowerCAmelCase : bool = True , _lowerCAmelCase : bool = True , ) -> Dict: """simple docstring""" __lowercase = self.get_auto_remove_tmp_dir("""./xxx""" , after=_lowerCAmelCase ) __lowercase = F'\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(_lowerCAmelCase )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n '.split() if fpaa: args.extend(["""--fp16"""] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __lowercase = F'--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json'.split() __lowercase = [F'{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py'] __lowercase = self.get_launcher(_lowerCAmelCase ) __lowercase = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_lowerCAmelCase , env=self.get_env() ) return output_dir def _a ( self : List[str] , _lowerCAmelCase : str=False ) -> Dict: """simple docstring""" __lowercase = min(2 , get_gpu_count() ) if distributed else 1 return F'deepspeed --num_nodes 1 --num_gpus {num_gpus}'.split()
80
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
1
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def snake_case ( lowerCamelCase = 8 ): '''simple docstring''' __lowercase = ascii_letters + digits + punctuation return "".join(secrets.choice(lowerCamelCase ) for _ in range(lowerCamelCase ) ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' i -= len(lowerCamelCase ) __lowercase = i // 3 __lowercase = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) __lowercase = ( chars_incl + random(lowerCamelCase , quotient + remainder ) + random(lowerCamelCase , lowerCamelCase ) + random(lowerCamelCase , lowerCamelCase ) ) __lowercase = list(lowerCamelCase ) shuffle(lowerCamelCase ) return "".join(lowerCamelCase ) # random is a generalised function for letters, characters and numbers def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' return "".join(secrets.choice(lowerCamelCase ) for _ in range(lowerCamelCase ) ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' pass # Put your code here... def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' pass # Put your code here... def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' pass # Put your code here... def snake_case ( lowerCamelCase , lowerCamelCase = 8 ): '''simple docstring''' if len(lowerCamelCase ) < min_length: # Your Password must be at least 8 characters long return False __lowercase = any(char in ascii_uppercase for char in password ) __lowercase = any(char in ascii_lowercase for char in password ) __lowercase = any(char in digits for char in password ) __lowercase = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def snake_case ( ): '''simple docstring''' __lowercase = int(input("""Please indicate the max length of your password: """ ).strip() ) __lowercase = input( """Please indicate the characters that must be in your password: """ ).strip() print("""Password generated:""" , password_generator(lowerCamelCase ) ) print( """Alternative Password generated:""" , alternative_password_generator(lowerCamelCase , lowerCamelCase ) , ) print("""[If you are thinking of using this passsword, You better save it.]""" ) if __name__ == "__main__": main()
80
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase : Dict = { """configuration_xlm_roberta_xl""": [ """XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaXLConfig""", """XLMRobertaXLOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Any = [ """XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaXLForCausalLM""", """XLMRobertaXLForMaskedLM""", """XLMRobertaXLForMultipleChoice""", """XLMRobertaXLForQuestionAnswering""", """XLMRobertaXLForSequenceClassification""", """XLMRobertaXLForTokenClassification""", """XLMRobertaXLModel""", """XLMRobertaXLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys __UpperCamelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
80
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' if not head: return True # split the list to two parts __lowercase , __lowercase = head.next, head while fast and fast.next: __lowercase = fast.next.next __lowercase = slow.next __lowercase = slow.next __lowercase = None # Don't forget here! But forget still works! # reverse the second part __lowercase = None while second: __lowercase = second.next __lowercase = node __lowercase = second __lowercase = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False __lowercase = node.next __lowercase = head.next return True def snake_case ( lowerCamelCase ): '''simple docstring''' if not head or not head.next: return True # 1. Get the midpoint (slow) __lowercase = __lowercase = __lowercase = head while fast and fast.next: __lowercase , __lowercase = fast.next.next, slow.next # 2. Push the second half into the stack __lowercase = [slow.val] while slow.next: __lowercase = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False __lowercase = cur.next return True def snake_case ( lowerCamelCase ): '''simple docstring''' if not head or not head.next: return True __lowercase = {} __lowercase = 0 while head: if head.val in d: d[head.val].append(lowerCamelCase ) else: __lowercase = [pos] __lowercase = head.next pos += 1 __lowercase = pos - 1 __lowercase = 0 for v in d.values(): if len(lowerCamelCase ) % 2 != 0: middle += 1 else: __lowercase = 0 for i in range(0 , len(lowerCamelCase ) ): if v[i] + v[len(lowerCamelCase ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
80
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
1
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : List[Any]=13 , _lowerCAmelCase : Union[str, Any]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[str]=99 , _lowerCAmelCase : str=64 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Dict=5 , _lowerCAmelCase : Optional[Any]=4 , _lowerCAmelCase : Union[str, Any]=37 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : Union[str, Any]=0.1 , _lowerCAmelCase : List[Any]=0.1 , _lowerCAmelCase : Dict=512 , _lowerCAmelCase : int=16 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : Any=3 , _lowerCAmelCase : str=4 , _lowerCAmelCase : Optional[int]=None , ) -> List[str]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = embedding_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : List[str] ) -> Any: """simple docstring""" return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def _a ( self : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any ) -> str: """simple docstring""" __lowercase = MegatronBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = MegatronBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Dict ) -> Any: """simple docstring""" __lowercase = MegatronBertForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = MegatronBertForNextSentencePrediction(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def _a ( self : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ) -> int: """simple docstring""" __lowercase = MegatronBertForPreTraining(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , next_sentence_label=_lowerCAmelCase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def _a ( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = MegatronBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = MegatronBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Dict , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Tuple: """simple docstring""" __lowercase = self.num_labels __lowercase = MegatronBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = MegatronBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Tuple = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) __snake_case :Optional[Any] = ( { 'feature-extraction': MegatronBertModel, 'fill-mask': MegatronBertForMaskedLM, 'question-answering': MegatronBertForQuestionAnswering, 'text-classification': MegatronBertForSequenceClassification, 'text-generation': MegatronBertForCausalLM, 'token-classification': MegatronBertForTokenClassification, 'zero-shot': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Dict = True # test_resize_embeddings = False __snake_case :Optional[int] = False def _a ( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str]=False ) -> str: """simple docstring""" __lowercase = super()._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) if return_labels: if model_class in get_values(_lowerCAmelCase ): __lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_lowerCAmelCase ) __lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) return inputs_dict def _a ( self : Dict ) -> str: """simple docstring""" __lowercase = MegatronBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def _a ( self : str ) -> Dict: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*_lowerCAmelCase ) def _a ( self : int ) -> Any: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : int ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*_lowerCAmelCase ) def _a ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*_lowerCAmelCase ) def _a ( self : Dict ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Dict ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*_lowerCAmelCase ) def snake_case ( lowerCamelCase ): '''simple docstring''' return torch.tensor( lowerCamelCase , dtype=torch.long , device=lowerCamelCase , ) __UpperCamelCase : List[str] = 1e-4 @require_torch @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): @slow @unittest.skip("""Model is not available.""" ) def _a ( self : List[str] ) -> Dict: """simple docstring""" __lowercase = """nvidia/megatron-bert-uncased-345m""" if "MYDIR" in os.environ: __lowercase = os.path.join(os.environ["""MYDIR"""] , _lowerCAmelCase ) __lowercase = MegatronBertModel.from_pretrained(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.half() __lowercase = _long_tensor([[101, 7110, 1005, 1056, 2023, 1_1333, 1_7413, 1029, 102]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase )[0] __lowercase = torch.Size((1, 9, 1024) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = [-0.6_040, -0.2_517, -0.1_025, 0.3_420, -0.6_758, -0.0_017, -0.1_089, -0.1_990, 0.5_728] for ii in range(3 ): for jj in range(3 ): __lowercase = output[0, ii, jj] __lowercase = expected[3 * ii + jj] __lowercase = """ii={} jj={} a={} b={}""".format(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) self.assertTrue(math.isclose(_lowerCAmelCase , _lowerCAmelCase , rel_tol=_lowerCAmelCase , abs_tol=_lowerCAmelCase ) , msg=_lowerCAmelCase )
80
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
1
from __future__ import annotations def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __lowercase , __lowercase = array[indexa], array[indexa] def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' if length > 1: __lowercase = int(length / 2 ) for i in range(lowerCamelCase , low + middle ): comp_and_swap(lowerCamelCase , lowerCamelCase , i + middle , lowerCamelCase ) bitonic_merge(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) bitonic_merge(lowerCamelCase , low + middle , lowerCamelCase , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' if length > 1: __lowercase = int(length / 2 ) bitonic_sort(lowerCamelCase , lowerCamelCase , lowerCamelCase , 1 ) bitonic_sort(lowerCamelCase , low + middle , lowerCamelCase , 0 ) bitonic_merge(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if __name__ == "__main__": __UpperCamelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : int = [int(item.strip()) for item in user_input.split(""",""")] bitonic_sort(unsorted, 0, len(unsorted), 1) print("""\nSorted array in ascending order is: """, end="""""") print(*unsorted, sep=""", """) bitonic_merge(unsorted, 0, len(unsorted), 0) print("""Sorted array in descending order is: """, end="""""") print(*unsorted, sep=""", """)
80
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
1
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __UpperCamelCase ( unittest.TestCase ): @property def _a ( self : List[Any] ) -> str: """simple docstring""" torch.manual_seed(0 ) __lowercase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = self.dummy_uncond_unet __lowercase = PNDMScheduler() __lowercase = PNDMPipeline(unet=_lowerCAmelCase , scheduler=_lowerCAmelCase ) pndm.to(_lowerCAmelCase ) pndm.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = torch.manual_seed(0 ) __lowercase = pndm(generator=_lowerCAmelCase , num_inference_steps=20 , output_type="""numpy""" ).images __lowercase = torch.manual_seed(0 ) __lowercase = pndm(generator=_lowerCAmelCase , num_inference_steps=20 , output_type="""numpy""" , return_dict=_lowerCAmelCase )[0] __lowercase = image[0, -3:, -3:, -1] __lowercase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowercase = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __UpperCamelCase ( unittest.TestCase ): def _a ( self : int ) -> List[str]: """simple docstring""" __lowercase = """google/ddpm-cifar10-32""" __lowercase = UNetaDModel.from_pretrained(_lowerCAmelCase ) __lowercase = PNDMScheduler() __lowercase = PNDMPipeline(unet=_lowerCAmelCase , scheduler=_lowerCAmelCase ) pndm.to(_lowerCAmelCase ) pndm.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = torch.manual_seed(0 ) __lowercase = pndm(generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowercase = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
80
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
1
def snake_case ( ): '''simple docstring''' for n in range(1 , 1_000_000 ): yield n * (n + 1) // 2 def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def snake_case ( ): '''simple docstring''' return next(i for i in triangle_number_generator() if count_divisors(lowerCamelCase ) > 500 ) if __name__ == "__main__": print(solution())
80
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
1
from bisect import bisect from itertools import accumulate def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = sorted(zip(lowerCamelCase , lowerCamelCase ) , key=lambda lowerCamelCase : x[0] / x[1] , reverse=lowerCamelCase ) __lowercase , __lowercase = [i[0] for i in r], [i[1] for i in r] __lowercase = list(accumulate(lowerCamelCase ) ) __lowercase = bisect(lowerCamelCase , lowerCamelCase ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
80
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase : Any = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Dict = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys __UpperCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
1
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Dict = 'char' __snake_case :List[Any] = 'bpe' __snake_case :int = 'wp' __UpperCamelCase : List[str] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :int = ['image_processor', 'char_tokenizer'] __snake_case :Dict = 'ViTImageProcessor' __snake_case :List[str] = 'MgpstrTokenizer' def __init__( self : Optional[int] , _lowerCAmelCase : List[str]=None , _lowerCAmelCase : int=None , **_lowerCAmelCase : Dict ) -> Dict: """simple docstring""" __lowercase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , _lowerCAmelCase , ) __lowercase = kwargs.pop("""feature_extractor""" ) __lowercase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) __lowercase = tokenizer __lowercase = AutoTokenizer.from_pretrained("""gpt2""" ) __lowercase = AutoTokenizer.from_pretrained("""bert-base-uncased""" ) super().__init__(_lowerCAmelCase , _lowerCAmelCase ) def __call__( self : str , _lowerCAmelCase : Optional[int]=None , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : List[str]=None , **_lowerCAmelCase : List[str] ) -> Dict: """simple docstring""" if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: __lowercase = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None: __lowercase = self.char_tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is None: return inputs elif images is None: return encodings else: __lowercase = encodings["""input_ids"""] return inputs def _a ( self : int , _lowerCAmelCase : int ) -> List[Any]: """simple docstring""" __lowercase , __lowercase , __lowercase = sequences __lowercase = char_preds.size(0 ) __lowercase , __lowercase = self._decode_helper(_lowerCAmelCase , """char""" ) __lowercase , __lowercase = self._decode_helper(_lowerCAmelCase , """bpe""" ) __lowercase , __lowercase = self._decode_helper(_lowerCAmelCase , """wp""" ) __lowercase = [] __lowercase = [] for i in range(_lowerCAmelCase ): __lowercase = [char_scores[i], bpe_scores[i], wp_scores[i]] __lowercase = [char_strs[i], bpe_strs[i], wp_strs[i]] __lowercase = scores.index(max(_lowerCAmelCase ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) __lowercase = {} __lowercase = final_strs __lowercase = final_scores __lowercase = char_strs __lowercase = bpe_strs __lowercase = wp_strs return out def _a ( self : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ) -> Tuple: """simple docstring""" if format == DecodeType.CHARACTER: __lowercase = self.char_decode __lowercase = 1 __lowercase = """[s]""" elif format == DecodeType.BPE: __lowercase = self.bpe_decode __lowercase = 2 __lowercase = """#""" elif format == DecodeType.WORDPIECE: __lowercase = self.wp_decode __lowercase = 102 __lowercase = """[SEP]""" else: raise ValueError(F'Format {format} is not supported.' ) __lowercase , __lowercase = [], [] __lowercase = pred_logits.size(0 ) __lowercase = pred_logits.size(1 ) __lowercase , __lowercase = pred_logits.topk(1 , dim=-1 , largest=_lowerCAmelCase , sorted=_lowerCAmelCase ) __lowercase = preds_index.view(-1 , _lowerCAmelCase )[:, 1:] __lowercase = decoder(_lowerCAmelCase ) __lowercase , __lowercase = torch.nn.functional.softmax(_lowerCAmelCase , dim=2 ).max(dim=2 ) __lowercase = preds_max_prob[:, 1:] for index in range(_lowerCAmelCase ): __lowercase = preds_str[index].find(_lowerCAmelCase ) __lowercase = preds_str[index][:pred_eos] __lowercase = preds_index[index].cpu().tolist() __lowercase = pred_index.index(_lowerCAmelCase ) if eos_token in pred_index else -1 __lowercase = preds_max_prob[index][: pred_eos_index + 1] __lowercase = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(_lowerCAmelCase ) conf_scores.append(_lowerCAmelCase ) return dec_strs, conf_scores def _a ( self : Optional[int] , _lowerCAmelCase : Any ) -> str: """simple docstring""" __lowercase = [seq.replace(""" """ , """""" ) for seq in self.char_tokenizer.batch_decode(_lowerCAmelCase )] return decode_strs def _a ( self : Union[str, Any] , _lowerCAmelCase : List[Any] ) -> int: """simple docstring""" return self.bpe_tokenizer.batch_decode(_lowerCAmelCase ) def _a ( self : str , _lowerCAmelCase : int ) -> Optional[Any]: """simple docstring""" __lowercase = [seq.replace(""" """ , """""" ) for seq in self.wp_tokenizer.batch_decode(_lowerCAmelCase )] return decode_strs
80
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
1
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand __UpperCamelCase : Optional[int] = ( """4S 3H 2C 7S 5H""", """9D 8H 2C 6S 7H""", """2D 6D 9D TH 7D""", """TC 8C 2S JH 6C""", """JH 8S TH AH QH""", """TS KS 5S 9S AC""", """KD 6S 9D TH AD""", """KS 8D 4D 9S 4S""", # pair """8C 4S KH JS 4D""", # pair """QH 8H KD JH 8S""", # pair """KC 4H KS 2H 8D""", # pair """KD 4S KC 3H 8S""", # pair """AH 8S AS KC JH""", # pair """3H 4C 4H 3S 2H""", # 2 pairs """5S 5D 2C KH KH""", # 2 pairs """3C KH 5D 5S KH""", # 2 pairs """AS 3C KH AD KH""", # 2 pairs """7C 7S 3S 7H 5S""", # 3 of a kind """7C 7S KH 2H 7H""", # 3 of a kind """AC KH QH AH AS""", # 3 of a kind """2H 4D 3C AS 5S""", # straight (low ace) """3C 5C 4C 2C 6H""", # straight """6S 8S 7S 5H 9H""", # straight """JS QS 9H TS KH""", # straight """QC KH TS JS AH""", # straight (high ace) """8C 9C 5C 3C TC""", # flush """3S 8S 9S 5S KS""", # flush """4C 5C 9C 8C KC""", # flush """JH 8H AH KH QH""", # flush """3D 2H 3H 2C 2D""", # full house """2H 2C 3S 3H 3D""", # full house """KH KC 3S 3H 3D""", # full house """JC 6H JS JD JH""", # 4 of a kind """JC 7H JS JD JH""", # 4 of a kind """JC KH JS JD JH""", # 4 of a kind """2S AS 4S 5S 3S""", # straight flush (low ace) """2D 6D 3D 4D 5D""", # straight flush """5C 6C 3C 7C 4C""", # straight flush """JH 9H TH KH QH""", # straight flush """JH AH TH KH QH""", # royal flush (high ace straight flush) ) __UpperCamelCase : str = ( ("""2H 3H 4H 5H 6H""", """KS AS TS QS JS""", """Loss"""), ("""2H 3H 4H 5H 6H""", """AS AD AC AH JD""", """Win"""), ("""AS AH 2H AD AC""", """JS JD JC JH 3D""", """Win"""), ("""2S AH 2H AS AC""", """JS JD JC JH AD""", """Loss"""), ("""2S AH 2H AS AC""", """2H 3H 5H 6H 7H""", """Win"""), ("""AS 3S 4S 8S 2S""", """2H 3H 5H 6H 7H""", """Win"""), ("""2H 3H 5H 6H 7H""", """2S 3H 4H 5S 6C""", """Win"""), ("""2S 3H 4H 5S 6C""", """3D 4C 5H 6H 2S""", """Tie"""), ("""2S 3H 4H 5S 6C""", """AH AC 5H 6H AS""", """Win"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H AS""", """Loss"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H 7S""", """Win"""), ("""6S AD 7H 4S AS""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S AH 4H 5S KC""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S 3H 6H 7S 9C""", """7H 3C TH 6H 9S""", """Loss"""), ("""4S 5H 6H TS AC""", """3S 5H 6H TS AC""", """Win"""), ("""2S AH 4H 5S 6C""", """AD 4C 5H 6H 2C""", """Tie"""), ("""AS AH 3H AD AC""", """AS AH 2H AD AC""", """Win"""), ("""AH AC 5H 5C QS""", """AH AC 5H 5C KS""", """Loss"""), ("""AH AC 5H 5C QS""", """KH KC 5H 5C QS""", """Win"""), ("""7C 7S KH 2H 7H""", """3C 3S AH 2H 3H""", """Win"""), ("""3C 3S AH 2H 3H""", """7C 7S KH 2H 7H""", """Loss"""), ("""6H 5H 4H 3H 2H""", """5H 4H 3H 2H AH""", """Win"""), ("""5H 4H 3H 2H AH""", """5H 4H 3H 2H AH""", """Tie"""), ("""5H 4H 3H 2H AH""", """6H 5H 4H 3H 2H""", """Loss"""), ("""AH AD KS KC AC""", """AH KD KH AC KC""", """Win"""), ("""2H 4D 3C AS 5S""", """2H 4D 3C 6S 5S""", """Loss"""), ("""2H 3S 3C 3H 2S""", """3S 3C 2S 2H 2D""", """Win"""), ("""4D 6D 5D 2D JH""", """3S 8S 3H TC KH""", """Loss"""), ("""4S 6C 8S 3S 7S""", """AD KS 2D 7D 7C""", """Loss"""), ("""6S 4C 7H 8C 3H""", """5H JC AH 9D 9C""", """Loss"""), ("""9D 9H JH TC QH""", """3C 2S JS 5C 7H""", """Win"""), ("""2H TC 8S AD 9S""", """4H TS 7H 2C 5C""", """Win"""), ("""9D 3S 2C 7S 7C""", """JC TD 3C TC 9H""", """Loss"""), ) __UpperCamelCase : Union[str, Any] = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", True), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", False), ("""AS 3S 4S 8S 2S""", True), ) __UpperCamelCase : List[Any] = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", False), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", True), ) __UpperCamelCase : List[str] = ( ("""2H 4D 3C AS 5S""", True, [5, 4, 3, 2, 14]), ("""2H 5D 3C AS 5S""", False, [14, 5, 5, 3, 2]), ("""JH QD KC AS TS""", False, [14, 13, 12, 11, 10]), ("""9D 3S 2C 7S 7C""", False, [9, 7, 7, 3, 2]), ) __UpperCamelCase : List[Any] = ( ("""JH AH TH KH QH""", 0), ("""JH 9H TH KH QH""", 0), ("""JC KH JS JD JH""", 7), ("""KH KC 3S 3H 3D""", 6), ("""8C 9C 5C 3C TC""", 0), ("""JS QS 9H TS KH""", 0), ("""7C 7S KH 2H 7H""", 3), ("""3C KH 5D 5S KH""", 2), ("""QH 8H KD JH 8S""", 1), ("""2D 6D 9D TH 7D""", 0), ) __UpperCamelCase : Any = ( ("""JH AH TH KH QH""", 23), ("""JH 9H TH KH QH""", 22), ("""JC KH JS JD JH""", 21), ("""KH KC 3S 3H 3D""", 20), ("""8C 9C 5C 3C TC""", 19), ("""JS QS 9H TS KH""", 18), ("""7C 7S KH 2H 7H""", 17), ("""3C KH 5D 5S KH""", 16), ("""QH 8H KD JH 8S""", 15), ("""2D 6D 9D TH 7D""", 14), ) def snake_case ( ): '''simple docstring''' __lowercase , __lowercase = randrange(len(lowerCamelCase ) ), randrange(len(lowerCamelCase ) ) __lowercase = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] __lowercase , __lowercase = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def snake_case ( lowerCamelCase = 100 ): '''simple docstring''' return (generate_random_hand() for _ in range(lowerCamelCase )) @pytest.mark.parametrize("""hand, expected""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = PokerHand(lowerCamelCase ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase ).compare_with(PokerHand(lowerCamelCase ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' assert PokerHand(lowerCamelCase ).compare_with(PokerHand(lowerCamelCase ) ) == expected def snake_case ( ): '''simple docstring''' __lowercase = [PokerHand(lowerCamelCase ) for hand in SORTED_HANDS] __lowercase = poker_hands.copy() shuffle(lowerCamelCase ) __lowercase = chain(sorted(lowerCamelCase ) ) for index, hand in enumerate(lowerCamelCase ): assert hand == poker_hands[index] def snake_case ( ): '''simple docstring''' __lowercase = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=lowerCamelCase ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def snake_case ( ): '''simple docstring''' __lowercase = PokerHand("""2C 4S AS 3D 5C""" ) __lowercase = True __lowercase = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def snake_case ( ): '''simple docstring''' __lowercase = 0 __lowercase = os.path.abspath(os.path.dirname(lowerCamelCase ) ) __lowercase = os.path.join(lowerCamelCase , """poker_hands.txt""" ) with open(lowerCamelCase ) as file_hand: for line in file_hand: __lowercase = line[:14].strip() __lowercase = line[15:].strip() __lowercase , __lowercase = PokerHand(lowerCamelCase ), PokerHand(lowerCamelCase ) __lowercase = player.compare_with(lowerCamelCase ) if output == "Win": answer += 1 assert answer == 376
80
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
1
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __UpperCamelCase : Optional[Any] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class __UpperCamelCase ( nn.Module ): def __init__( self : int , _lowerCAmelCase : Dict ) -> Optional[Any]: """simple docstring""" super().__init__() __lowercase = torchvision.models.resnetaaa(pretrained=_lowerCAmelCase ) __lowercase = list(model.children() )[:-2] __lowercase = nn.Sequential(*_lowerCAmelCase ) __lowercase = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def _a ( self : str , _lowerCAmelCase : List[str] ) -> int: """simple docstring""" __lowercase = self.pool(self.model(_lowerCAmelCase ) ) __lowercase = torch.flatten(_lowerCAmelCase , start_dim=2 ) __lowercase = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = [json.loads(_lowerCAmelCase ) for l in open(_lowerCAmelCase )] __lowercase = os.path.dirname(_lowerCAmelCase ) __lowercase = tokenizer __lowercase = labels __lowercase = len(_lowerCAmelCase ) __lowercase = max_seq_length __lowercase = transforms def __len__( self : int ) -> Tuple: """simple docstring""" return len(self.data ) def __getitem__( self : Any , _lowerCAmelCase : Any ) -> str: """simple docstring""" __lowercase = torch.LongTensor(self.tokenizer.encode(self.data[index]["""text"""] , add_special_tokens=_lowerCAmelCase ) ) __lowercase , __lowercase , __lowercase = sentence[0], sentence[1:-1], sentence[-1] __lowercase = sentence[: self.max_seq_length] __lowercase = torch.zeros(self.n_classes ) __lowercase = 1 __lowercase = Image.open(os.path.join(self.data_dir , self.data[index]["""img"""] ) ).convert("""RGB""" ) __lowercase = self.transforms(_lowerCAmelCase ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def _a ( self : List[Any] ) -> Any: """simple docstring""" __lowercase = Counter() for row in self.data: label_freqs.update(row["""label"""] ) return label_freqs def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = [len(row["""sentence"""] ) for row in batch] __lowercase , __lowercase = len(lowerCamelCase ), max(lowerCamelCase ) __lowercase = torch.zeros(lowerCamelCase , lowerCamelCase , dtype=torch.long ) __lowercase = torch.zeros(lowerCamelCase , lowerCamelCase , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(lowerCamelCase , lowerCamelCase ) ): __lowercase = input_row["""sentence"""] __lowercase = 1 __lowercase = torch.stack([row["""image"""] for row in batch] ) __lowercase = torch.stack([row["""label"""] for row in batch] ) __lowercase = torch.stack([row["""image_start_token"""] for row in batch] ) __lowercase = torch.stack([row["""image_end_token"""] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def snake_case ( ): '''simple docstring''' return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def snake_case ( ): '''simple docstring''' return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.46777044, 0.44531429, 0.40661017] , std=[0.12221994, 0.12145835, 0.14380469] , ), ] )
80
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' if len(lowerCamelCase ) <= 1: return [tuple(lowerCamelCase )] __lowercase = [] def generate(lowerCamelCase , lowerCamelCase ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , lowerCamelCase ) for i in range(k - 1 ): if k % 2 == 0: # k is even __lowercase , __lowercase = arr[k - 1], arr[i] else: # k is odd __lowercase , __lowercase = arr[k - 1], arr[0] generate(k - 1 , lowerCamelCase ) generate(len(lowerCamelCase ) , lowerCamelCase ) return res if __name__ == "__main__": __UpperCamelCase : List[str] = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : int = [int(item) for item in user_input.split(""",""")] print(heaps(arr))
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
1
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument __UpperCamelCase : Optional[int] = { """/attention/""": """/0/SelfAttention/""", """/self_attention/""": """/0/SelfAttention/""", """/encoder_decoder_attention/""": """/1/EncDecAttention/""", """value""": """v""", """query""": """q""", """key""": """k""", """out""": """o""", """pre_self_attention_layer_norm""": """0/layer_norm""", """pre_cross_attention_layer_norm""": """1/layer_norm""", """pre_attention_layer_norm""": """0/layer_norm""", # previously 1, but seems wrong """token_embedder""": """shared""", """encoder_norm""": """final_layer_norm""", """decoder_norm""": """final_layer_norm""", """relpos_bias/rel_embedding""": """block/0/layer/0/SelfAttention/relative_attention_bias/weight""", """router/router_weights/w/""": """router/classifier/""", """roer/roer_weights/w/""": """router/classifier/""", """logits_dense""": """lm_head""", } def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = list(s_dict.keys() ) for key in keys: __lowercase = r""".*/layers_(\d+)""" __lowercase = key if re.match(lowerCamelCase , lowerCamelCase ): __lowercase = re.sub(r"""layers_(\d+)""" , r"""block/\1/layer""" , lowerCamelCase ) __lowercase = r"""(encoder|decoder)\/""" if re.match(lowerCamelCase , lowerCamelCase ): __lowercase = re.match(lowerCamelCase , lowerCamelCase ).groups() if groups[0] == "encoder": __lowercase = re.sub(r"""/mlp/""" , r"""/1/mlp/""" , lowerCamelCase ) __lowercase = re.sub(r"""/pre_mlp_layer_norm/""" , r"""/1/layer_norm/""" , lowerCamelCase ) elif groups[0] == "decoder": __lowercase = re.sub(r"""/mlp/""" , r"""/2/mlp/""" , lowerCamelCase ) __lowercase = re.sub(r"""/pre_mlp_layer_norm/""" , r"""/2/layer_norm/""" , lowerCamelCase ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: __lowercase = new_key.replace(lowerCamelCase , lowerCamelCase ) print(F'{key} -> {new_key}' ) __lowercase = s_dict.pop(lowerCamelCase ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __lowercase = s_dict[ """encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight""" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __lowercase = s_dict[ """decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight""" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: __lowercase = s_dict[key].shape[0] __lowercase = s_dict[key] for idx in range(lowerCamelCase ): __lowercase = expert_weihts[idx] print(F'{key} -> {key.replace("expert/" , "nested fstring" )}' ) s_dict.pop(lowerCamelCase ) return s_dict __UpperCamelCase : int = { """NUM_ENCODER_LAYERS""": """num_layers""", """NUM_DECODER_LAYERS""": """num_decoder_layers""", """NUM_HEADS""": """num_heads""", """HEAD_DIM""": """d_kv""", """EMBED_DIM""": """d_model""", """MLP_DIM""": """d_ff""", """NUM_SELECTED_EXPERTS""": """num_selected_experts""", """NUM_ENCODER_SPARSE_LAYERS""": """num_sparse_encoder_layers""", """NUM_DECODER_SPARSE_LAYERS""": """num_sparse_decoder_layers""", """dense.MlpBlock.activations""": """feed_forward_proj""", } def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' import regex as re with open(lowerCamelCase , """r""" ) as f: __lowercase = f.read() __lowercase = re.findall(r"""(.*) = ([0-9.]*)""" , lowerCamelCase ) __lowercase = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": __lowercase = float(lowerCamelCase ) if """.""" in value else int(lowerCamelCase ) __lowercase = re.findall(r"""(.*activations) = \(\'(.*)\',\)""" , lowerCamelCase )[0] __lowercase = str(activation[1] ) __lowercase = num_experts __lowercase = SwitchTransformersConfig(**lowerCamelCase ) return config def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase="./" , lowerCamelCase=8 ): '''simple docstring''' print(F'Loading flax weights from : {flax_checkpoint_path}' ) __lowercase = checkpoints.load_tax_checkpoint(lowerCamelCase ) if gin_file is not None: __lowercase = convert_gin_to_config(lowerCamelCase , lowerCamelCase ) else: __lowercase = SwitchTransformersConfig.from_pretrained(lowerCamelCase ) __lowercase = SwitchTransformersForConditionalGeneration(lowerCamelCase ) __lowercase = flax_params["""target"""] __lowercase = flatten_dict(lowerCamelCase , sep="""/""" ) __lowercase = rename_keys(lowerCamelCase ) __lowercase = unflatten_dict(lowerCamelCase , sep="""/""" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(lowerCamelCase , lowerCamelCase ) print(F'Save PyTorch model to {pytorch_dump_path}' ) pt_model.save_pretrained(lowerCamelCase ) if __name__ == "__main__": __UpperCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the""" """ model architecture. If not provided, a `gin_file` has to be provided.""" ), ) parser.add_argument( """--gin_file""", default=None, type=str, required=False, help="""Path to the gin config file. If not provided, a `config_file` has to be passed """, ) parser.add_argument( """--config_name""", default=None, type=str, required=False, help="""Config name of SwitchTransformers model.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output pytorch model.""" ) parser.add_argument("""--num_experts""", default=8, type=int, required=False, help="""Number of experts""") __UpperCamelCase : int = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
80
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
1
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCamelCase : def __init__( self : int , _lowerCAmelCase : Dict , _lowerCAmelCase : int=13 , _lowerCAmelCase : Optional[int]=7 , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Union[str, Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : Any=True , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : List[str]=False , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : Optional[Any]=False , _lowerCAmelCase : Any=2 , _lowerCAmelCase : Union[str, Any]=99 , _lowerCAmelCase : int=0 , _lowerCAmelCase : Tuple=32 , _lowerCAmelCase : Union[str, Any]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : List[Any]=512 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : List[str]=2 , _lowerCAmelCase : Optional[int]=4 , _lowerCAmelCase : Dict="last" , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : int=None , _lowerCAmelCase : str=0 , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_lengths __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = gelu_activation __lowercase = sinusoidal_embeddings __lowercase = causal __lowercase = asm __lowercase = n_langs __lowercase = vocab_size __lowercase = n_special __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = summary_type __lowercase = use_proj __lowercase = scope __lowercase = bos_token_id def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_input_lengths: __lowercase = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , 2 ).float() __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _a ( self : List[Any] ) -> str: """simple docstring""" return XLMConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , num_labels=self.num_labels , bos_token_id=self.bos_token_id , ) def _a ( self : Dict , _lowerCAmelCase : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict , ) -> int: """simple docstring""" __lowercase = XLMModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , lengths=_lowerCAmelCase , langs=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , langs=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str] , ) -> List[Any]: """simple docstring""" __lowercase = XLMWithLMHeadModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : str , ) -> str: """simple docstring""" __lowercase = XLMForQuestionAnsweringSimple(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) __lowercase = outputs self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , ) -> List[str]: """simple docstring""" __lowercase = XLMForQuestionAnswering(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = model( _lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , cls_index=_lowerCAmelCase , is_impossible=_lowerCAmelCase , p_mask=_lowerCAmelCase , ) __lowercase = model( _lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , cls_index=_lowerCAmelCase , is_impossible=_lowerCAmelCase , ) ((__lowercase) , ) = result_with_labels.to_tuple() __lowercase = model(_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) ((__lowercase) , ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _a ( self : Optional[int] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any , ) -> Dict: """simple docstring""" __lowercase = XLMForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] , ) -> Tuple: """simple docstring""" __lowercase = self.num_labels __lowercase = XLMForTokenClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict , _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , ) -> Optional[int]: """simple docstring""" __lowercase = self.num_choices __lowercase = XLMForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """lengths""": input_lengths} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Dict = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) __snake_case :Tuple = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable __snake_case :Optional[Any] = ( { 'feature-extraction': XLMModel, 'fill-mask': XLMWithLMHeadModel, 'question-answering': XLMForQuestionAnsweringSimple, 'text-classification': XLMForSequenceClassification, 'text-generation': XLMWithLMHeadModel, 'token-classification': XLMForTokenClassification, 'zero-shot': XLMForSequenceClassification, } if is_torch_available() else {} ) def _a ( self : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> int: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _a ( self : Tuple , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any]=False ) -> Union[str, Any]: """simple docstring""" __lowercase = super()._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": __lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) __lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) return inputs_dict def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = XLMModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , emb_dim=37 ) def _a ( self : List[Any] ) -> Any: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*_lowerCAmelCase ) def _a ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*_lowerCAmelCase ) def _a ( self : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any=False , _lowerCAmelCase : List[str]=1 ) -> Tuple: """simple docstring""" self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual( [isinstance(_lowerCAmelCase , _lowerCAmelCase ) for iter_attentions in attentions] , [True] * len(_lowerCAmelCase ) ) self.assertEqual(len(_lowerCAmelCase ) , (max_length - min_length) * num_beam_groups ) for idx, iter_attentions in enumerate(_lowerCAmelCase ): # adds PAD dummy token __lowercase = min_length + idx + 1 __lowercase = min_length + idx + 1 __lowercase = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions] , [expected_shape] * len(_lowerCAmelCase ) ) def _a ( self : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]=False , _lowerCAmelCase : Dict=1 ) -> int: """simple docstring""" self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual( [isinstance(_lowerCAmelCase , _lowerCAmelCase ) for iter_hidden_states in hidden_states] , [True] * len(_lowerCAmelCase ) , ) self.assertEqual(len(_lowerCAmelCase ) , (max_length - min_length) * num_beam_groups ) for idx, iter_hidden_states in enumerate(_lowerCAmelCase ): # adds PAD dummy token __lowercase = min_length + idx + 1 __lowercase = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states] , [expected_shape] * len(_lowerCAmelCase ) , ) pass @slow def _a ( self : Tuple ) -> Optional[Any]: """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = XLMModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = XLMWithLMHeadModel.from_pretrained("""xlm-mlm-en-2048""" ) model.to(_lowerCAmelCase ) __lowercase = torch.tensor([[14, 447]] , dtype=torch.long , device=_lowerCAmelCase ) # the president __lowercase = [ 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, 14, 447, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference __lowercase = model.generate(_lowerCAmelCase , do_sample=_lowerCAmelCase ) self.assertListEqual(output_ids[0].cpu().numpy().tolist() , _lowerCAmelCase )
80
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
1
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = FileLock(str(tmpdir / """foo.lock""" ) ) __lowercase = FileLock(str(tmpdir / """foo.lock""" ) ) __lowercase = 0.01 with locka.acquire(): with pytest.raises(lowerCamelCase ): __lowercase = time.time() locka.acquire(lowerCamelCase ) assert time.time() - _start > timeout def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = """a""" * 1_000 + """.lock""" __lowercase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(""".lock""" ) assert not locka._lock_file.endswith(lowerCamelCase ) assert len(os.path.basename(locka._lock_file ) ) <= 255 __lowercase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(lowerCamelCase ): locka.acquire(0 )
80
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Dict = 'naver-clova-ix/donut-base-finetuned-docvqa' __snake_case :Dict = ( 'This is a tool that answers a question about an document (pdf). It takes an input named `document` which ' 'should be the document containing the information, as well as a `question` that is the question about the ' 'document. It returns a text that contains the answer to the question.' ) __snake_case :List[str] = 'document_qa' __snake_case :List[Any] = AutoProcessor __snake_case :str = VisionEncoderDecoderModel __snake_case :List[Any] = ['image', 'text'] __snake_case :Optional[Any] = ['text'] def __init__( self : Any , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : Any ) -> Optional[int]: """simple docstring""" if not is_vision_available(): raise ValueError("""Pillow must be installed to use the DocumentQuestionAnsweringTool.""" ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) def _a ( self : Any , _lowerCAmelCase : "Image" , _lowerCAmelCase : str ) -> str: """simple docstring""" __lowercase = """<s_docvqa><s_question>{user_input}</s_question><s_answer>""" __lowercase = task_prompt.replace("""{user_input}""" , _lowerCAmelCase ) __lowercase = self.pre_processor.tokenizer( _lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_tensors="""pt""" ).input_ids __lowercase = self.pre_processor(_lowerCAmelCase , return_tensors="""pt""" ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _a ( self : List[Any] , _lowerCAmelCase : Dict ) -> str: """simple docstring""" return self.model.generate( inputs["""pixel_values"""].to(self.device ) , decoder_input_ids=inputs["""decoder_input_ids"""].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=_lowerCAmelCase , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=_lowerCAmelCase , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=_lowerCAmelCase , ).sequences def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.pre_processor.batch_decode(_lowerCAmelCase )[0] __lowercase = sequence.replace(self.pre_processor.tokenizer.eos_token , """""" ) __lowercase = sequence.replace(self.pre_processor.tokenizer.pad_token , """""" ) __lowercase = re.sub(r"""<.*?>""" , """""" , _lowerCAmelCase , count=1 ).strip() # remove first task start token __lowercase = self.pre_processor.tokenajson(_lowerCAmelCase ) return sequence["answer"]
80
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
1
from math import isclose, sqrt def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = point_y / 4 / point_x __lowercase = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) __lowercase = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) __lowercase = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 __lowercase = outgoing_gradient**2 + 4 __lowercase = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) __lowercase = (point_y - outgoing_gradient * point_x) ** 2 - 100 __lowercase = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) __lowercase = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point __lowercase = x_minus if isclose(lowerCamelCase , lowerCamelCase ) else x_plus __lowercase = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def snake_case ( lowerCamelCase = 1.4 , lowerCamelCase = -9.6 ): '''simple docstring''' __lowercase = 0 __lowercase = first_x_coord __lowercase = first_y_coord __lowercase = (10.1 - point_y) / (0.0 - point_x) while not (-0.01 <= point_x <= 0.01 and point_y > 0): __lowercase , __lowercase , __lowercase = next_point(lowerCamelCase , lowerCamelCase , lowerCamelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F'''{solution() = }''')
80
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
1
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
1
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version __UpperCamelCase : Optional[int] = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize __UpperCamelCase : Union[str, Any] = """\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } """ __UpperCamelCase : List[str] = """\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. """ __UpperCamelCase : Optional[int] = """ Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: 'meteor': meteor score. Examples: >>> meteor = datasets.load_metric('meteor') >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"] >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results[\"meteor\"], 4)) 0.6944 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"""] , reference_urls=[ """https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score""", """https://en.wikipedia.org/wiki/METEOR""", ] , ) def _a ( self : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" import nltk nltk.download("""wordnet""" ) if NLTK_VERSION >= version.Version("""3.6.5""" ): nltk.download("""punkt""" ) if NLTK_VERSION >= version.Version("""3.6.6""" ): nltk.download("""omw-1.4""" ) def _a ( self : List[str] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int]=0.9 , _lowerCAmelCase : Optional[Any]=3 , _lowerCAmelCase : Tuple=0.5 ) -> Optional[Any]: """simple docstring""" if NLTK_VERSION >= version.Version("""3.6.5""" ): __lowercase = [ meteor_score.single_meteor_score( word_tokenize(_lowerCAmelCase ) , word_tokenize(_lowerCAmelCase ) , alpha=_lowerCAmelCase , beta=_lowerCAmelCase , gamma=_lowerCAmelCase ) for ref, pred in zip(_lowerCAmelCase , _lowerCAmelCase ) ] else: __lowercase = [ meteor_score.single_meteor_score(_lowerCAmelCase , _lowerCAmelCase , alpha=_lowerCAmelCase , beta=_lowerCAmelCase , gamma=_lowerCAmelCase ) for ref, pred in zip(_lowerCAmelCase , _lowerCAmelCase ) ] return {"meteor": np.mean(_lowerCAmelCase )}
80
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class __UpperCamelCase : def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any]=14 , _lowerCAmelCase : Optional[int]=7 , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : Tuple=99 , _lowerCAmelCase : int=32 , _lowerCAmelCase : str=5 , _lowerCAmelCase : Optional[Any]=4 , _lowerCAmelCase : Optional[Any]=37 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : Union[str, Any]=0.1 , _lowerCAmelCase : List[Any]=0.1 , _lowerCAmelCase : int=512 , _lowerCAmelCase : str=16 , _lowerCAmelCase : int=2 , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : str=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : str=None , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_token_type_ids __lowercase = use_input_mask __lowercase = use_labels __lowercase = use_mc_token_ids __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = self.vocab_size - 1 def _a ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = None if self.use_mc_token_ids: __lowercase = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() __lowercase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def _a ( self : List[Any] ) -> List[str]: """simple docstring""" return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , *_lowerCAmelCase : List[str] ) -> List[Any]: """simple docstring""" __lowercase = CTRLModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , head_mask=_lowerCAmelCase ) model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def _a ( self : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , *_lowerCAmelCase : List[str] ) -> List[str]: """simple docstring""" __lowercase = CTRLLMHeadModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : Optional[Any] ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def _a ( self : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , *_lowerCAmelCase : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.num_labels __lowercase = CTRLForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () __snake_case :Optional[int] = (CTRLLMHeadModel,) if is_torch_available() else () __snake_case :List[Any] = ( { 'feature-extraction': CTRLModel, 'text-classification': CTRLForSequenceClassification, 'text-generation': CTRLLMHeadModel, 'zero-shot': CTRLForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Dict = True __snake_case :Union[str, Any] = False __snake_case :str = False def _a ( self : str , _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> Dict: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = CTRLModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , n_embd=37 ) def _a ( self : Dict ) -> Optional[int]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def _a ( self : str ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*_lowerCAmelCase ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _a ( self : List[Any] ) -> str: """simple docstring""" pass @slow def _a ( self : List[Any] ) -> str: """simple docstring""" for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CTRLModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def _a ( self : int ) -> List[Any]: """simple docstring""" pass @require_torch class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Dict ) -> Any: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(_lowerCAmelCase ) __lowercase = torch.tensor( [[1_1859, 0, 1611, 8]] , dtype=torch.long , device=_lowerCAmelCase ) # Legal the president is __lowercase = [ 1_1859, 0, 1611, 8, 5, 150, 2_6449, 2, 19, 348, 469, 3, 2595, 48, 2_0740, 24_6533, 24_6533, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __lowercase = model.generate(_lowerCAmelCase , do_sample=_lowerCAmelCase ) self.assertListEqual(output_ids[0].tolist() , _lowerCAmelCase )
80
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase : Optional[int] = { """configuration_rembert""": ["""REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RemBertConfig""", """RemBertOnnxConfig"""] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[str] = ["""RemBertTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Optional[int] = ["""RemBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : str = [ """REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """RemBertForCausalLM""", """RemBertForMaskedLM""", """RemBertForMultipleChoice""", """RemBertForQuestionAnswering""", """RemBertForSequenceClassification""", """RemBertForTokenClassification""", """RemBertLayer""", """RemBertModel""", """RemBertPreTrainedModel""", """load_tf_weights_in_rembert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRemBertForCausalLM""", """TFRemBertForMaskedLM""", """TFRemBertForMultipleChoice""", """TFRemBertForQuestionAnswering""", """TFRemBertForSequenceClassification""", """TFRemBertForTokenClassification""", """TFRemBertLayer""", """TFRemBertModel""", """TFRemBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("""Program to check whether a number is a Perfect number or not...""") __UpperCamelCase : Dict = int(input("""Enter number: """).strip()) print(F'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
80
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
1
__UpperCamelCase : int = tuple[float, float, float] __UpperCamelCase : Optional[int] = tuple[float, float, float] def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = end_pointa[0] - end_pointa[0] __lowercase = end_pointa[1] - end_pointa[1] __lowercase = end_pointa[2] - end_pointa[2] return (x, y, z) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = ab[1] * ac[2] - ab[2] * ac[1] # *i __lowercase = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j __lowercase = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' return tuple(round(lowerCamelCase , lowerCamelCase ) for x in vector ) == (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = 10 ): '''simple docstring''' __lowercase = create_vector(lowerCamelCase , lowerCamelCase ) __lowercase = create_vector(lowerCamelCase , lowerCamelCase ) return is_zero_vector(get_ad_vectors_cross(lowerCamelCase , lowerCamelCase ) , lowerCamelCase )
80
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
1
import re def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = re.compile( r"""^(?:0|94|\+94|0{2}94)""" r"""7(0|1|2|4|5|6|7|8)""" r"""(-| |)""" r"""\d{7}$""" ) return bool(re.search(lowerCamelCase , lowerCamelCase ) ) if __name__ == "__main__": __UpperCamelCase : Dict = """0094702343221""" print(is_sri_lankan_phone_number(phone))
80
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
1
from __future__ import annotations def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = sorted(numsa + numsa ) __lowercase , __lowercase = divmod(len(lowerCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Optional[int] = [float(x) for x in input("""Enter the elements of first array: """).split()] __UpperCamelCase : Optional[int] = [float(x) for x in input("""Enter the elements of second array: """).split()] print(F'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
80
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
1
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
1
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCamelCase ( _lowerCAmelCase ): def _a ( self : Union[str, Any] , _lowerCAmelCase : float ) -> float: """simple docstring""" return 0.0 def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) __lowercase = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = 512 __lowercase = [1] + [0] * (size - 1) __lowercase = [filter_type.process(lowerCamelCase ) for item in inputs] __lowercase = [0] * (samplerate - size) # zero-padding outputs += filler __lowercase = np.abs(np.fft.fft(lowerCamelCase ) ) __lowercase = 20 * np.logaa(lowerCamelCase ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("""Frequency (Hz)""" ) plt.xscale("""log""" ) # Display within reasonable bounds __lowercase = get_bounds(lowerCamelCase , lowerCamelCase ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel("""Gain (dB)""" ) plt.plot(lowerCamelCase ) plt.show() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = 512 __lowercase = [1] + [0] * (size - 1) __lowercase = [filter_type.process(lowerCamelCase ) for item in inputs] __lowercase = [0] * (samplerate - size) # zero-padding outputs += filler __lowercase = np.angle(np.fft.fft(lowerCamelCase ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("""Frequency (Hz)""" ) plt.xscale("""log""" ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel("""Phase shift (Radians)""" ) plt.plot(np.unwrap(lowerCamelCase , -2 * pi ) ) plt.show()
80
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
1
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP __UpperCamelCase : List[Any] = False try: __UpperCamelCase : int = _is_package_available("""google.colab""") except ModuleNotFoundError: pass @input.register class __UpperCamelCase : def __init__( self : Dict , _lowerCAmelCase : str = None , _lowerCAmelCase : list = [] ) -> Tuple: """simple docstring""" __lowercase = 0 __lowercase = choices __lowercase = prompt if sys.platform == "win32": __lowercase = """*""" else: __lowercase = """➔ """ def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str = "" ) -> str: """simple docstring""" if sys.platform != "win32": writeColor(self.choices[index] , 32 , _lowerCAmelCase ) else: forceWrite(self.choices[index] , _lowerCAmelCase ) def _a ( self : Dict , _lowerCAmelCase : int ) -> Any: """simple docstring""" if index == self.position: forceWrite(F' {self.arrow_char} ' ) self.write_choice(_lowerCAmelCase ) else: forceWrite(F' {self.choices[index]}' ) reset_cursor() def _a ( self : Optional[Any] , _lowerCAmelCase : Direction , _lowerCAmelCase : int = 1 ) -> Dict: """simple docstring""" __lowercase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(_lowerCAmelCase ) move_cursor(_lowerCAmelCase , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["""up"""] ) def _a ( self : Any ) -> Dict: """simple docstring""" self.move_direction(Direction.UP ) @input.mark(KEYMAP["""down"""] ) def _a ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["""newline"""] ) def _a ( self : Tuple ) -> Dict: """simple docstring""" move_cursor(len(self.choices ) - self.position , """DOWN""" ) return self.position @input.mark(KEYMAP["""interrupt"""] ) def _a ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" move_cursor(len(self.choices ) - self.position , """DOWN""" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(_lowerCAmelCase )] for number in range(10 )] ) def _a ( self : Union[str, Any] ) -> int: """simple docstring""" __lowercase = int(chr(self.current_selection ) ) __lowercase = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , _lowerCAmelCase ) else: return else: return def _a ( self : Optional[int] , _lowerCAmelCase : int = 0 ) -> Union[str, Any]: """simple docstring""" if self.prompt: linebreak() forceWrite(self.prompt , """\n""" ) if in_colab: forceWrite("""Please input a choice index (starting from 0), and press enter""" , """\n""" ) else: forceWrite("""Please select a choice using the arrow or number keys, and selecting with enter""" , """\n""" ) __lowercase = default_choice for i in range(len(self.choices ) ): self.print_choice(_lowerCAmelCase ) forceWrite("""\n""" ) move_cursor(len(self.choices ) - self.position , """UP""" ) with cursor.hide(): while True: if in_colab: try: __lowercase = int(builtins.input() ) except ValueError: __lowercase = default_choice else: __lowercase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , """UP""" ) clear_line() self.write_choice(_lowerCAmelCase , """\n""" ) return choice
80
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
1
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder __UpperCamelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name __UpperCamelCase : List[Any] = 256 class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[str] = ['melgan'] def __init__( self : Dict , _lowerCAmelCase : SpectrogramNotesEncoder , _lowerCAmelCase : SpectrogramContEncoder , _lowerCAmelCase : TaFilmDecoder , _lowerCAmelCase : DDPMScheduler , _lowerCAmelCase : OnnxRuntimeModel if is_onnx_available() else Any , ) -> None: """simple docstring""" super().__init__() # From MELGAN __lowercase = math.log(1e-5 ) # Matches MelGAN training. __lowercase = 4.0 # Largest value for most examples __lowercase = 128 self.register_modules( notes_encoder=_lowerCAmelCase , continuous_encoder=_lowerCAmelCase , decoder=_lowerCAmelCase , scheduler=_lowerCAmelCase , melgan=_lowerCAmelCase , ) def _a ( self : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple=(-1.0, 1.0) , _lowerCAmelCase : Optional[Any]=False ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = output_range if clip: __lowercase = torch.clip(_lowerCAmelCase , self.min_value , self.max_value ) # Scale to [0, 1]. __lowercase = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def _a ( self : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any]=(-1.0, 1.0) , _lowerCAmelCase : Tuple=False ) -> List[str]: """simple docstring""" __lowercase , __lowercase = input_range __lowercase = torch.clip(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if clip else outputs # Scale to [0, 1]. __lowercase = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def _a ( self : str , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str ) -> Optional[int]: """simple docstring""" __lowercase = input_tokens > 0 __lowercase , __lowercase = self.notes_encoder( encoder_input_tokens=_lowerCAmelCase , encoder_inputs_mask=_lowerCAmelCase ) __lowercase , __lowercase = self.continuous_encoder( encoder_inputs=_lowerCAmelCase , encoder_inputs_mask=_lowerCAmelCase ) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def _a ( self : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] ) -> Dict: """simple docstring""" __lowercase = noise_time if not torch.is_tensor(_lowerCAmelCase ): __lowercase = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device ) elif torch.is_tensor(_lowerCAmelCase ) and len(timesteps.shape ) == 0: __lowercase = timesteps[None].to(input_tokens.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __lowercase = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device ) __lowercase = self.decoder( encodings_and_masks=_lowerCAmelCase , decoder_input_tokens=_lowerCAmelCase , decoder_noise_time=_lowerCAmelCase ) return logits @torch.no_grad() def __call__( self : Tuple , _lowerCAmelCase : List[List[int]] , _lowerCAmelCase : Optional[torch.Generator] = None , _lowerCAmelCase : int = 100 , _lowerCAmelCase : bool = True , _lowerCAmelCase : str = "numpy" , _lowerCAmelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowerCAmelCase : int = 1 , ) -> Union[AudioPipelineOutput, Tuple]: """simple docstring""" if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or callback_steps <= 0) ): raise ValueError( F'`callback_steps` has to be a positive integer but is {callback_steps} of type' F' {type(_lowerCAmelCase )}.' ) __lowercase = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa ) __lowercase = np.zeros([1, 0, self.n_dims] , np.floataa ) __lowercase = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=_lowerCAmelCase , device=self.device ) for i, encoder_input_tokens in enumerate(_lowerCAmelCase ): if i == 0: __lowercase = torch.from_numpy(pred_mel[:1].copy() ).to( device=self.device , dtype=self.decoder.dtype ) # The first chunk has no previous context. __lowercase = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=_lowerCAmelCase , device=self.device ) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. __lowercase = ones __lowercase = self.scale_features( _lowerCAmelCase , output_range=[-1.0, 1.0] , clip=_lowerCAmelCase ) __lowercase = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens] ).to(device=self.device ) , continuous_inputs=_lowerCAmelCase , continuous_mask=_lowerCAmelCase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop __lowercase = randn_tensor( shape=encoder_continuous_inputs.shape , generator=_lowerCAmelCase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(_lowerCAmelCase ) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): __lowercase = self.decode( encodings_and_masks=_lowerCAmelCase , input_tokens=_lowerCAmelCase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 __lowercase = self.scheduler.step(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , generator=_lowerCAmelCase ).prev_sample __lowercase = self.scale_to_features(_lowerCAmelCase , input_range=[-1.0, 1.0] ) __lowercase = mel[:1] __lowercase = mel.cpu().float().numpy() __lowercase = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1 ) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_lowerCAmelCase , _lowerCAmelCase ) logger.info("""Generated segment""" , _lowerCAmelCase ) if output_type == "numpy" and not is_onnx_available(): raise ValueError( """Cannot return output in 'np' format if ONNX is not available. Make sure to have ONNX installed or set 'output_type' to 'mel'.""" ) elif output_type == "numpy" and self.melgan is None: raise ValueError( """Cannot return output in 'np' format if melgan component is not defined. Make sure to define `self.melgan` or set 'output_type' to 'mel'.""" ) if output_type == "numpy": __lowercase = self.melgan(input_features=full_pred_mel.astype(np.floataa ) ) else: __lowercase = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=_lowerCAmelCase )
80
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
1
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
1
class __UpperCamelCase : # Public class to implement a graph def __init__( self : str , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : list[list[bool]] ) -> None: """simple docstring""" __lowercase = row __lowercase = col __lowercase = graph def _a ( self : str , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : list[list[bool]] ) -> bool: """simple docstring""" return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def _a ( self : str , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : list[list[bool]] ) -> None: """simple docstring""" __lowercase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowercase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowercase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , _lowerCAmelCase ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , _lowerCAmelCase ) def _a ( self : Dict ) -> int: # And finally, count all islands. """simple docstring""" __lowercase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowercase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) count += 1 return count
80
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
1
import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("""1.0.0a"""): raise Exception("""requires fairseq >= 1.0.0a""") logging.set_verbosity_info() __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : str = """Hello world! cécé herlolip""" def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = FairseqRobertaModel.from_pretrained(lowerCamelCase ) roberta.eval() # disable dropout __lowercase = roberta.model.encoder.sentence_encoder __lowercase = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1e-5 , ) if classification_head: __lowercase = roberta.model.classification_heads["""mnli"""].out_proj.weight.shape[0] print("""Our RoBERTa config:""" , lowerCamelCase ) __lowercase = XLMRobertaXLForSequenceClassification(lowerCamelCase ) if classification_head else XLMRobertaXLForMaskedLM(lowerCamelCase ) model.eval() # Now let's copy all the weights. # Embeddings __lowercase = roberta_sent_encoder.embed_tokens.weight __lowercase = roberta_sent_encoder.embed_positions.weight __lowercase = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. __lowercase = roberta_sent_encoder.layer_norm.weight __lowercase = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer __lowercase = model.roberta.encoder.layer[i] __lowercase = roberta_sent_encoder.layers[i] __lowercase = layer.attention __lowercase = roberta_layer.self_attn_layer_norm.weight __lowercase = roberta_layer.self_attn_layer_norm.bias # self attention __lowercase = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) __lowercase = roberta_layer.self_attn.q_proj.weight __lowercase = roberta_layer.self_attn.q_proj.bias __lowercase = roberta_layer.self_attn.k_proj.weight __lowercase = roberta_layer.self_attn.k_proj.bias __lowercase = roberta_layer.self_attn.v_proj.weight __lowercase = roberta_layer.self_attn.v_proj.bias # self-attention output __lowercase = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape __lowercase = roberta_layer.self_attn.out_proj.weight __lowercase = roberta_layer.self_attn.out_proj.bias # this one is final layer norm __lowercase = roberta_layer.final_layer_norm.weight __lowercase = roberta_layer.final_layer_norm.bias # intermediate __lowercase = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape __lowercase = roberta_layer.fca.weight __lowercase = roberta_layer.fca.bias # output __lowercase = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape __lowercase = roberta_layer.fca.weight __lowercase = roberta_layer.fca.bias # end of layer if classification_head: __lowercase = roberta.model.classification_heads["""mnli"""].dense.weight __lowercase = roberta.model.classification_heads["""mnli"""].dense.bias __lowercase = roberta.model.classification_heads["""mnli"""].out_proj.weight __lowercase = roberta.model.classification_heads["""mnli"""].out_proj.bias else: # LM Head __lowercase = roberta.model.encoder.lm_head.dense.weight __lowercase = roberta.model.encoder.lm_head.dense.bias __lowercase = roberta.model.encoder.lm_head.layer_norm.weight __lowercase = roberta.model.encoder.lm_head.layer_norm.bias __lowercase = roberta.model.encoder.lm_head.weight __lowercase = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. __lowercase = roberta.encode(lowerCamelCase ).unsqueeze(0 ) # batch of size 1 __lowercase = model(lowerCamelCase )[0] if classification_head: __lowercase = roberta.model.classification_heads["""mnli"""](roberta.extract_features(lowerCamelCase ) ) else: __lowercase = roberta.model(lowerCamelCase )[0] print(our_output.shape , their_output.shape ) __lowercase = torch.max(torch.abs(our_output - their_output ) ).item() print(F'max_absolute_diff = {max_absolute_diff}' ) # ~ 1e-7 __lowercase = torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) print("""Do both models output the same tensors?""" , """🔥""" if success else """💩""" ) if not success: raise Exception("""Something went wRoNg""" ) pathlib.Path(lowerCamelCase ).mkdir(parents=lowerCamelCase , exist_ok=lowerCamelCase ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(lowerCamelCase ) if __name__ == "__main__": __UpperCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--roberta_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--classification_head""", action="""store_true""", help="""Whether to convert a final classification head.""" ) __UpperCamelCase : List[Any] = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
80
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' if not all(x.isalpha() for x in string ): raise ValueError("""String must only contain alphabetic characters.""" ) __lowercase = sorted(string.lower() ) return len(lowerCamelCase ) == len(set(lowerCamelCase ) ) if __name__ == "__main__": __UpperCamelCase : Dict = input("""Enter a string """).strip() __UpperCamelCase : Optional[Any] = is_isogram(input_str) print(F'''{input_str} is {"an" if isogram else "not an"} isogram.''')
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
1
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class __UpperCamelCase ( pl.LightningModule ): def __init__( self : Any , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" super().__init__() __lowercase = model __lowercase = 2 __lowercase = nn.Linear(self.model.config.hidden_size , self.num_labels ) def _a ( self : int ) -> List[str]: """simple docstring""" pass def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = LongformerModel.from_pretrained(lowerCamelCase ) __lowercase = LightningModel(lowerCamelCase ) __lowercase = torch.load(lowerCamelCase , map_location=torch.device("""cpu""" ) ) lightning_model.load_state_dict(ckpt["""state_dict"""] ) # init longformer question answering model __lowercase = LongformerForQuestionAnswering.from_pretrained(lowerCamelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(lowerCamelCase ) print(F'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": __UpperCamelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __UpperCamelCase : List[str] = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
80
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __UpperCamelCase : Optional[Any] = { """configuration_convnext""": ["""CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvNextConfig""", """ConvNextOnnxConfig"""] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Optional[Any] = ["""ConvNextFeatureExtractor"""] __UpperCamelCase : List[str] = ["""ConvNextImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvNextForImageClassification""", """ConvNextModel""", """ConvNextPreTrainedModel""", """ConvNextBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Union[str, Any] = [ """TFConvNextForImageClassification""", """TFConvNextModel""", """TFConvNextPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys __UpperCamelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
80
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
1
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) __UpperCamelCase : Tuple = logging.getLogger() __UpperCamelCase : str = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __UpperCamelCase ( _lowerCAmelCase ): def _a ( self : Tuple , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) __lowercase = {"""source""": """What is love ?""", """target""": """life"""} __lowercase = {"""train""": 12, """val""": 2, """test""": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: __lowercase = """\n""".join([contents[field]] * n_lines[split] ) with open(os.path.join(_lowerCAmelCase , F'{split}.{field}' ) , """w""" ) as f: f.write(_lowerCAmelCase ) def _a ( self : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : str = "pytorch" ) -> int: """simple docstring""" __lowercase = self.get_auto_remove_tmp_dir() __lowercase = os.path.join(_lowerCAmelCase , """output""" ) __lowercase = os.path.join(_lowerCAmelCase , """data""" ) self._create_dummy_data(data_dir=_lowerCAmelCase ) __lowercase = F'\n --data_dir {data_dir} \\n --output_dir {output_dir} \\n --model_name_or_path facebook/rag-sequence-base \\n --model_type rag_sequence \\n --do_train \\n --do_predict \\n --n_val -1 \\n --val_check_interval 1.0 \\n --train_batch_size 2 \\n --eval_batch_size 1 \\n --max_source_length 25 \\n --max_target_length 25 \\n --val_max_target_length 25 \\n --test_max_target_length 25 \\n --label_smoothing 0.1 \\n --dropout 0.1 \\n --attention_dropout 0.1 \\n --weight_decay 0.001 \\n --adam_epsilon 1e-08 \\n --max_grad_norm 0.1 \\n --lr_scheduler polynomial \\n --learning_rate 3e-04 \\n --num_train_epochs 1 \\n --warmup_steps 4 \\n --gradient_accumulation_steps 1 \\n --distributed-port 8787 \\n --use_dummy_dataset 1 \\n --distributed_retriever {distributed_retriever} \\n '.split() if gpus > 0: testargs.append(F'--gpus={gpus}' ) if is_apex_available(): testargs.append("""--fp16""" ) else: testargs.append("""--gpus=0""" ) testargs.append("""--distributed_backend=ddp_cpu""" ) testargs.append("""--num_processes=2""" ) __lowercase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(_lowerCAmelCase , env=self.get_env() ) __lowercase = os.path.join(_lowerCAmelCase , """metrics.json""" ) with open(_lowerCAmelCase ) as f: __lowercase = json.load(_lowerCAmelCase ) return result @require_torch_gpu def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_gpu @require_ray def _a ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu @require_ray def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 )
80
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
1
import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase : Optional[Any] = get_tests_dir("""fixtures/test_sentencepiece_bpe_char.model""") @require_sentencepiece @require_tokenizers class __UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): __snake_case :Tuple = SpeechTaTokenizer __snake_case :int = False __snake_case :Dict = True def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowercase = SpeechTaTokenizer(_lowerCAmelCase ) __lowercase = AddedToken("""<mask>""" , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) __lowercase = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def _a ( self : Optional[int] , _lowerCAmelCase : Optional[Any] ) -> Any: """simple docstring""" __lowercase = """this is a test""" __lowercase = """this is a test""" return input_text, output_text def _a ( self : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[Any]=False , _lowerCAmelCase : str=20 , _lowerCAmelCase : List[str]=5 ) -> str: """simple docstring""" __lowercase , __lowercase = self.get_input_output_texts(_lowerCAmelCase ) __lowercase = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) __lowercase = tokenizer.decode(_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) return text, ids def _a ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = """<pad>""" __lowercase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def _a ( self : str ) -> Any: """simple docstring""" __lowercase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-4] , """œ""" ) self.assertEqual(vocab_keys[-2] , """<mask>""" ) self.assertEqual(vocab_keys[-1] , """<ctc_blank>""" ) self.assertEqual(len(_lowerCAmelCase ) , 81 ) def _a ( self : List[Any] ) -> str: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def _a ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=_lowerCAmelCase ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): __lowercase = tokenizer.vocab_size __lowercase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(_lowerCAmelCase ) __lowercase = tokenizer.vocab_size __lowercase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , len(_lowerCAmelCase ) ) self.assertEqual(_lowerCAmelCase , all_size + len(_lowerCAmelCase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=_lowerCAmelCase ) self.assertGreaterEqual(len(_lowerCAmelCase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(_lowerCAmelCase ) __lowercase = tokenizer.vocab_size __lowercase = len(_lowerCAmelCase ) self.assertNotEqual(_lowerCAmelCase , 0 ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , len(_lowerCAmelCase ) ) self.assertEqual(_lowerCAmelCase , all_size_a + len(_lowerCAmelCase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=_lowerCAmelCase ) self.assertGreaterEqual(len(_lowerCAmelCase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Optional[Any] ) -> Any: """simple docstring""" __lowercase = self.get_tokenizer() __lowercase = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(_lowerCAmelCase , [SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) __lowercase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( _lowerCAmelCase , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) __lowercase = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) # fmt: off self.assertListEqual(_lowerCAmelCase , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on __lowercase = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def _a ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off __lowercase = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name="""microsoft/speecht5_asr""" , revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" , sequences=_lowerCAmelCase , )
80
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
1
from __future__ import annotations from math import ceil, floor, sqrt def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0] __lowercase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowercase = 0 # the area corresponding to the grid that gives the product closest to target __lowercase = 0 # an estimate of b, using the quadratic formula __lowercase = 42 # the largest integer less than b_estimate __lowercase = 42 # the largest integer less than b_estimate __lowercase = 42 # the triangle number corresponding to b_floor __lowercase = 42 # the triangle number corresponding to b_ceil __lowercase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowercase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowercase = floor(lowerCamelCase ) __lowercase = ceil(lowerCamelCase ) __lowercase = triangle_numbers[b_floor] __lowercase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowercase = triangle_b_first_guess * triangle_a __lowercase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowercase = triangle_b_second_guess * triangle_a __lowercase = idx_a * b_ceil return area if __name__ == "__main__": print(F'''{solution() = }''')
80
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
1
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class __UpperCamelCase : def __init__( self : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : Any=13 , _lowerCAmelCase : str=7 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Any=True , _lowerCAmelCase : Union[str, Any]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : str=32 , _lowerCAmelCase : Dict=2 , _lowerCAmelCase : Union[str, Any]=4 , _lowerCAmelCase : Any=37 , _lowerCAmelCase : Union[str, Any]="gelu" , _lowerCAmelCase : Union[str, Any]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : str=512 , _lowerCAmelCase : Optional[int]=16 , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : str=3 , _lowerCAmelCase : Tuple=4 , _lowerCAmelCase : List[str]=None , _lowerCAmelCase : str=1000 , ) -> Dict: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = range_bbox def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment __lowercase = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __lowercase = bbox[i, j, 3] __lowercase = bbox[i, j, 1] __lowercase = t if bbox[i, j, 2] < bbox[i, j, 0]: __lowercase = bbox[i, j, 2] __lowercase = bbox[i, j, 0] __lowercase = t __lowercase = tf.convert_to_tensor(_lowerCAmelCase ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] ) -> Tuple: """simple docstring""" __lowercase = TFLayoutLMModel(config=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a ( self : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = TFLayoutLMForMaskedLM(config=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = TFLayoutLMForSequenceClassification(config=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : Dict ) -> Dict: """simple docstring""" __lowercase = self.num_labels __lowercase = TFLayoutLMForTokenClassification(config=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : str , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = TFLayoutLMForQuestionAnswering(config=_lowerCAmelCase ) __lowercase = model(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = { """input_ids""": input_ids, """bbox""": bbox, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_tf class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :List[Any] = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __snake_case :Union[str, Any] = ( { 'feature-extraction': TFLayoutLMModel, 'fill-mask': TFLayoutLMForMaskedLM, 'text-classification': TFLayoutLMForSequenceClassification, 'token-classification': TFLayoutLMForTokenClassification, 'zero-shot': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __snake_case :str = False __snake_case :Dict = True __snake_case :Tuple = 1_0 def _a ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = TFLayoutLMModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Any ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) @slow def _a ( self : Any ) -> Any: """simple docstring""" for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = TFLayoutLMModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip("""Onnx compliancy broke with TF 2.10""" ) def _a ( self : str ) -> Any: """simple docstring""" pass def snake_case ( ): '''simple docstring''' __lowercase = tf.convert_to_tensor([[101,1_019,1_014,1_016,1_037,12_849,4_747,1_004,14_246,2_278,5_439,4_524,5_002,2_930,2_193,2_930,4_341,3_208,1_005,1_055,2_171,2_848,11_300,3_531,102],[101,4_070,4_034,7_020,1_024,3_058,1_015,1_013,2_861,1_013,6_070,19_274,2_772,6_205,27_814,16_147,16_147,4_343,2_047,10_283,10_969,14_389,1_012,2_338,102]] ) # noqa: E231 __lowercase = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 __lowercase = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1_000,1_000,1_000,1_000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1_000,1_000,1_000,1_000]]] ) # noqa: E231 __lowercase = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) __lowercase = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : str ) -> str: """simple docstring""" __lowercase = TFLayoutLMModel.from_pretrained("""microsoft/layoutlm-base-uncased""" ) __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = prepare_layoutlm_batch_inputs() # forward pass __lowercase = model(input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) # test the sequence output on [0, :3, :3] __lowercase = tf.convert_to_tensor( [[0.1_785, -0.1_947, -0.0_425], [-0.3_254, -0.2_807, 0.2_553], [-0.5_391, -0.3_322, 0.3_364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , _lowerCAmelCase , atol=1e-3 ) ) # test the pooled output on [1, :3] __lowercase = tf.convert_to_tensor([-0.6_580, -0.0_214, 0.8_552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , _lowerCAmelCase , atol=1e-3 ) ) @slow def _a ( self : int ) -> Any: """simple docstring""" __lowercase = TFLayoutLMForSequenceClassification.from_pretrained("""microsoft/layoutlm-base-uncased""" , num_labels=2 ) __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = prepare_layoutlm_batch_inputs() # forward pass __lowercase = model( input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar __lowercase = outputs.loss __lowercase = (2,) self.assertEqual(loss.shape , _lowerCAmelCase ) # test the shape of the logits __lowercase = outputs.logits __lowercase = (2, 2) self.assertEqual(logits.shape , _lowerCAmelCase ) @slow def _a ( self : List[str] ) -> int: """simple docstring""" __lowercase = TFLayoutLMForTokenClassification.from_pretrained("""microsoft/layoutlm-base-uncased""" , num_labels=13 ) __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = prepare_layoutlm_batch_inputs() # forward pass __lowercase = model( input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) # test the shape of the logits __lowercase = outputs.logits __lowercase = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , _lowerCAmelCase ) @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = TFLayoutLMForQuestionAnswering.from_pretrained("""microsoft/layoutlm-base-uncased""" ) __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = prepare_layoutlm_batch_inputs() # forward pass __lowercase = model(input_ids=_lowerCAmelCase , bbox=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) # test the shape of the logits __lowercase = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , _lowerCAmelCase ) self.assertEqual(outputs.end_logits.shape , _lowerCAmelCase )
80
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
1
from functools import lru_cache def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = 2 __lowercase = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(lowerCamelCase ) if n > 1: factors.add(lowerCamelCase ) return factors @lru_cache def snake_case ( lowerCamelCase ): '''simple docstring''' return len(unique_prime_factors(lowerCamelCase ) ) def snake_case ( lowerCamelCase ): '''simple docstring''' return len(set(lowerCamelCase ) ) in (0, 1) def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = 2 while True: # Increment each value of a generated range __lowercase = [base + i for i in range(lowerCamelCase )] # Run elements through out unique_prime_factors function # Append our target number to the end. __lowercase = [upf_len(lowerCamelCase ) for x in group] checker.append(lowerCamelCase ) # If all numbers in the list are equal, return the group variable. if equality(lowerCamelCase ): return group # Increment our base variable by 1 base += 1 def snake_case ( lowerCamelCase = 4 ): '''simple docstring''' __lowercase = run(lowerCamelCase ) return results[0] if len(lowerCamelCase ) else None if __name__ == "__main__": print(solution())
80
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
1
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
1
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = fname.split(os.path.sep )[-1] return re.search(r"""^(.*)_\d+\.jpg$""" , lowerCamelCase ).groups()[0] class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple=None , _lowerCAmelCase : str=None ) -> Any: """simple docstring""" __lowercase = file_names __lowercase = image_transform __lowercase = label_to_id def __len__( self : Tuple ) -> Optional[int]: """simple docstring""" return len(self.file_names ) def __getitem__( self : List[Any] , _lowerCAmelCase : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.file_names[idx] __lowercase = PIL.Image.open(_lowerCAmelCase ) __lowercase = raw_image.convert("""RGB""" ) if self.image_transform is not None: __lowercase = self.image_transform(_lowerCAmelCase ) __lowercase = extract_label(_lowerCAmelCase ) if self.label_to_id is not None: __lowercase = self.label_to_id[label] return {"image": image, "label": label} def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if args.with_tracking: __lowercase = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="""all""" , project_dir=args.project_dir ) else: __lowercase = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __lowercase = config["""lr"""] __lowercase = int(config["""num_epochs"""] ) __lowercase = int(config["""seed"""] ) __lowercase = int(config["""batch_size"""] ) __lowercase = config["""image_size"""] if not isinstance(lowerCamelCase , (list, tuple) ): __lowercase = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , """isdigit""" ): if args.checkpointing_steps == "epoch": __lowercase = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): __lowercase = int(args.checkpointing_steps ) else: raise ValueError( F'Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.' ) else: __lowercase = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: __lowercase = os.path.split(lowerCamelCase )[-1].split(""".""" )[0] accelerator.init_trackers(lowerCamelCase , lowerCamelCase ) # Grab all the image filenames __lowercase = [os.path.join(args.data_dir , lowerCamelCase ) for fname in os.listdir(args.data_dir ) if fname.endswith(""".jpg""" )] # Build the label correspondences __lowercase = [extract_label(lowerCamelCase ) for fname in file_names] __lowercase = list(set(lowerCamelCase ) ) id_to_label.sort() __lowercase = {lbl: i for i, lbl in enumerate(lowerCamelCase )} # Set the seed before splitting the data. np.random.seed(lowerCamelCase ) torch.manual_seed(lowerCamelCase ) torch.cuda.manual_seed_all(lowerCamelCase ) # Split our filenames between train and validation __lowercase = np.random.permutation(len(lowerCamelCase ) ) __lowercase = int(0.8 * len(lowerCamelCase ) ) __lowercase = random_perm[:cut] __lowercase = random_perm[cut:] # For training we use a simple RandomResizedCrop __lowercase = Compose([RandomResizedCrop(lowerCamelCase , scale=(0.5, 1.0) ), ToTensor()] ) __lowercase = PetsDataset( [file_names[i] for i in train_split] , image_transform=lowerCamelCase , label_to_id=lowerCamelCase ) # For evaluation, we use a deterministic Resize __lowercase = Compose([Resize(lowerCamelCase ), ToTensor()] ) __lowercase = PetsDataset([file_names[i] for i in eval_split] , image_transform=lowerCamelCase , label_to_id=lowerCamelCase ) # Instantiate dataloaders. __lowercase = DataLoader(lowerCamelCase , shuffle=lowerCamelCase , batch_size=lowerCamelCase , num_workers=4 ) __lowercase = DataLoader(lowerCamelCase , shuffle=lowerCamelCase , batch_size=lowerCamelCase , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __lowercase = create_model("""resnet50d""" , pretrained=lowerCamelCase , num_classes=len(lowerCamelCase ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __lowercase = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): __lowercase = False for param in model.get_classifier().parameters(): __lowercase = True # We normalize the batches of images to be a bit faster. __lowercase = torch.tensor(model.default_cfg["""mean"""] )[None, :, None, None].to(accelerator.device ) __lowercase = torch.tensor(model.default_cfg["""std"""] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer __lowercase = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler __lowercase = OneCycleLR(optimizer=lowerCamelCase , max_lr=lowerCamelCase , epochs=lowerCamelCase , steps_per_epoch=len(lowerCamelCase ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # We need to keep track of how many total steps we have iterated over __lowercase = 0 # We also need to keep track of the starting epoch so files are named properly __lowercase = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(F'Resumed from checkpoint: {args.resume_from_checkpoint}' ) accelerator.load_state(args.resume_from_checkpoint ) __lowercase = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint __lowercase = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) __lowercase = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` __lowercase = os.path.splitext(lowerCamelCase )[0] if "epoch" in training_difference: __lowercase = int(training_difference.replace("""epoch_""" , """""" ) ) + 1 __lowercase = None else: __lowercase = int(training_difference.replace("""step_""" , """""" ) ) __lowercase = resume_step // len(lowerCamelCase ) resume_step -= starting_epoch * len(lowerCamelCase ) # Now we train the model for epoch in range(lowerCamelCase , lowerCamelCase ): model.train() if args.with_tracking: __lowercase = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step __lowercase = accelerator.skip_first_batches(lowerCamelCase , lowerCamelCase ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader __lowercase = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. __lowercase = {k: v.to(accelerator.device ) for k, v in batch.items()} __lowercase = (batch["""image"""] - mean) / std __lowercase = model(lowerCamelCase ) __lowercase = torch.nn.functional.cross_entropy(lowerCamelCase , batch["""label"""] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(lowerCamelCase , lowerCamelCase ): __lowercase = F'step_{overall_step}' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: __lowercase = os.path.join(args.output_dir , lowerCamelCase ) accelerator.save_state(lowerCamelCase ) model.eval() __lowercase = 0 __lowercase = 0 for step, batch in enumerate(lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. __lowercase = {k: v.to(accelerator.device ) for k, v in batch.items()} __lowercase = (batch["""image"""] - mean) / std with torch.no_grad(): __lowercase = model(lowerCamelCase ) __lowercase = outputs.argmax(dim=-1 ) __lowercase , __lowercase = accelerator.gather_for_metrics((predictions, batch["""label"""]) ) __lowercase = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() __lowercase = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}: {100 * eval_metric:.2f}' ) if args.with_tracking: accelerator.log( { """accuracy""": 100 * eval_metric, """train_loss""": total_loss.item() / len(lowerCamelCase ), """epoch""": epoch, } , step=lowerCamelCase , ) if checkpointing_steps == "epoch": __lowercase = F'epoch_{epoch}' if args.output_dir is not None: __lowercase = os.path.join(args.output_dir , lowerCamelCase ) accelerator.save_state(lowerCamelCase ) if args.with_tracking: accelerator.end_training() def snake_case ( ): '''simple docstring''' __lowercase = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument("""--data_dir""" , required=lowerCamelCase , help="""The data folder on disk.""" ) parser.add_argument("""--fp16""" , action="""store_true""" , help="""If passed, will use FP16 training.""" ) parser.add_argument( """--mixed_precision""" , type=lowerCamelCase , default=lowerCamelCase , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) parser.add_argument( """--checkpointing_steps""" , type=lowerCamelCase , default=lowerCamelCase , help="""Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch.""" , ) parser.add_argument( """--output_dir""" , type=lowerCamelCase , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--resume_from_checkpoint""" , type=lowerCamelCase , default=lowerCamelCase , help="""If the training should continue from a checkpoint folder.""" , ) parser.add_argument( """--with_tracking""" , action="""store_true""" , help="""Whether to load in all available experiment trackers from the environment and use them for logging.""" , ) parser.add_argument( """--project_dir""" , type=lowerCamelCase , default="""logs""" , help="""Location on where to store experiment tracking logs` and relevent project information""" , ) __lowercase = parser.parse_args() __lowercase = {"""lr""": 3e-2, """num_epochs""": 3, """seed""": 42, """batch_size""": 64, """image_size""": 224} training_function(lowerCamelCase , lowerCamelCase ) if __name__ == "__main__": main()
80
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : int = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Union[str, Any] = 'lxmert' __snake_case :Union[str, Any] = {} def __init__( self : List[str] , _lowerCAmelCase : Dict=3_0522 , _lowerCAmelCase : List[str]=768 , _lowerCAmelCase : Union[str, Any]=12 , _lowerCAmelCase : Union[str, Any]=9500 , _lowerCAmelCase : Union[str, Any]=1600 , _lowerCAmelCase : Optional[Any]=400 , _lowerCAmelCase : Tuple=3072 , _lowerCAmelCase : List[Any]="gelu" , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=512 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : List[str]=1e-12 , _lowerCAmelCase : Any=9 , _lowerCAmelCase : Optional[Any]=5 , _lowerCAmelCase : Any=5 , _lowerCAmelCase : Dict=2048 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[Any]=6.67 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=True , _lowerCAmelCase : int=True , **_lowerCAmelCase : Tuple , ) -> Dict: """simple docstring""" __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = num_qa_labels __lowercase = num_object_labels __lowercase = num_attr_labels __lowercase = l_layers __lowercase = x_layers __lowercase = r_layers __lowercase = visual_feat_dim __lowercase = visual_pos_dim __lowercase = visual_loss_normalizer __lowercase = task_matched __lowercase = task_mask_lm __lowercase = task_obj_predict __lowercase = task_qa __lowercase = visual_obj_loss __lowercase = visual_attr_loss __lowercase = visual_feat_loss __lowercase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**_lowerCAmelCase )
80
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Tuple = { """configuration_groupvit""": [ """GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GroupViTConfig""", """GroupViTOnnxConfig""", """GroupViTTextConfig""", """GroupViTVisionConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : int = [ """GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GroupViTModel""", """GroupViTPreTrainedModel""", """GroupViTTextModel""", """GroupViTVisionModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFGroupViTModel""", """TFGroupViTPreTrainedModel""", """TFGroupViTTextModel""", """TFGroupViTVisionModel""", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys __UpperCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : int , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=99 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Any=4 , _lowerCAmelCase : Tuple=37 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : List[str]=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Union[str, Any]=512 , _lowerCAmelCase : Dict=16 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : int=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Optional[Any] ) -> int: """simple docstring""" return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _a ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = DistilBertModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = DistilBertForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> str: """simple docstring""" __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __snake_case :Dict = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case :Tuple = True __snake_case :Tuple = True __snake_case :List[str] = True __snake_case :Optional[int] = True def _a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , dim=37 ) def _a ( self : Dict ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCAmelCase ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCAmelCase ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCAmelCase ) def _a ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCAmelCase ) @slow def _a ( self : int ) -> Optional[Any]: """simple docstring""" for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @slow @require_torch_gpu def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = torch.jit.trace( _lowerCAmelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , """traced_model.pt""" ) ) __lowercase = torch.jit.load(os.path.join(_lowerCAmelCase , """traced_model.pt""" ) , map_location=_lowerCAmelCase ) loaded(inputs_dict["""input_ids"""].to(_lowerCAmelCase ) , inputs_dict["""attention_mask"""].to(_lowerCAmelCase ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = DistilBertModel.from_pretrained("""distilbert-base-uncased""" ) __lowercase = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] __lowercase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _lowerCAmelCase , atol=1e-4 ) )
80
1
from __future__ import annotations class __UpperCamelCase : def __init__( self : Dict , _lowerCAmelCase : int ) -> None: """simple docstring""" __lowercase = data __lowercase = None __lowercase = None def snake_case ( lowerCamelCase ): # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def snake_case ( lowerCamelCase ): '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def snake_case ( lowerCamelCase ): '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def snake_case ( ): # Main function for testing. '''simple docstring''' __lowercase = Node(1 ) __lowercase = Node(2 ) __lowercase = Node(3 ) __lowercase = Node(4 ) __lowercase = Node(5 ) __lowercase = Node(6 ) __lowercase = Node(7 ) __lowercase = Node(8 ) __lowercase = Node(9 ) print(is_full_binary_tree(lowerCamelCase ) ) print(depth_of_tree(lowerCamelCase ) ) print("""Tree is: """ ) display(lowerCamelCase ) if __name__ == "__main__": main()
80
import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class __UpperCamelCase ( _lowerCAmelCase ): # to overwrite at feature extractactor specific tests __snake_case :Optional[int] = None __snake_case :Dict = None @property def _a ( self : str ) -> List[str]: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """feature_size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """padding_value""" ) ) def _a ( self : Tuple ) -> str: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) for x, y in zip(_lowerCAmelCase , processed_features[input_name] ) ) ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _a ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_lowerCAmelCase ) __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowercase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowercase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _a ( self : str , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : int ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Dict , _lowerCAmelCase : Tuple ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = self.feat_extract_tester.seq_length_diff __lowercase = self.feat_extract_tester.max_seq_length + pad_diff __lowercase = self.feat_extract_tester.min_seq_length __lowercase = self.feat_extract_tester.batch_size __lowercase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , padding=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) __lowercase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" )[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowercase = feat_extract.pad(_lowerCAmelCase , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , pad_to_multiple_of=10 ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] self.assertTrue(all(len(_lowerCAmelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_lowerCAmelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowercase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def _a ( self : Tuple , _lowerCAmelCase : str=False ) -> Union[str, Any]: """simple docstring""" def _inputs_have_equal_length(_lowerCAmelCase : Tuple ): __lowercase = len(input[0] ) for input_slice in input[1:]: if len(_lowerCAmelCase ) != length: return False return True def _inputs_are_equal(_lowerCAmelCase : Any , _lowerCAmelCase : str ): if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False for input_slice_a, input_slice_a in zip(_lowerCAmelCase , _lowerCAmelCase ): if not np.allclose(np.asarray(_lowerCAmelCase ) , np.asarray(_lowerCAmelCase ) , atol=1e-3 ): return False return True __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_lowerCAmelCase ) __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to smallest with np __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) # truncate to middle __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase , return_tensors="""np""" , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_lowerCAmelCase ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowercase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(_inputs_are_equal(_lowerCAmelCase , _lowerCAmelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""longest""" , truncation=_lowerCAmelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_lowerCAmelCase ): feat_extract.pad(_lowerCAmelCase , padding="""max_length""" , truncation=_lowerCAmelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowercase = 12 __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , truncation=_lowerCAmelCase , ) __lowercase = input_a[input_name] __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_lowerCAmelCase , ) __lowercase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowercase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowercase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_lowerCAmelCase ) ) self.assertFalse(_inputs_have_equal_length(_lowerCAmelCase ) ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" self._check_padding(numpify=_lowerCAmelCase ) def _a ( self : int ) -> Tuple: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) def _a ( self : str ) -> str: """simple docstring""" self._check_truncation(numpify=_lowerCAmelCase ) @require_torch def _a ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def _a ( self : Any ) -> Any: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_dict ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" )[input_name] __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def _a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = feat_extract.pad(_lowerCAmelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _lowerCAmelCase ) def _a ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.feat_extract_dict __lowercase = True __lowercase = self.feature_extraction_class(**_lowerCAmelCase ) __lowercase = self.feat_extract_tester.prepare_inputs_for_common() __lowercase = [len(_lowerCAmelCase ) for x in speech_inputs] __lowercase = feat_extract.model_input_names[0] __lowercase = BatchFeature({input_name: speech_inputs} ) __lowercase = min(_lowerCAmelCase ) __lowercase = feat_extract.pad( _lowerCAmelCase , padding="""max_length""" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
80
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : List[str] , _lowerCAmelCase : pyspark.sql.DataFrame , _lowerCAmelCase : Optional[NamedSplit] = None , _lowerCAmelCase : Optional[Features] = None , _lowerCAmelCase : bool = True , _lowerCAmelCase : str = None , _lowerCAmelCase : bool = False , _lowerCAmelCase : str = None , _lowerCAmelCase : bool = True , _lowerCAmelCase : str = "arrow" , **_lowerCAmelCase : List[str] , ) -> Optional[Any]: """simple docstring""" super().__init__( split=_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase , streaming=_lowerCAmelCase , **_lowerCAmelCase , ) __lowercase = load_from_cache_file __lowercase = file_format __lowercase = Spark( df=_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase , working_dir=_lowerCAmelCase , **_lowerCAmelCase , ) def _a ( self : Dict ) -> str: """simple docstring""" if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) __lowercase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=_lowerCAmelCase , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
80
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [[] for _ in range(lowerCamelCase )] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1 or len(lowerCamelCase ) <= key: return input_string for position, character in enumerate(lowerCamelCase ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowerCamelCase ) __lowercase = ["""""".join(lowerCamelCase ) for row in temp_grid] __lowercase = """""".join(lowerCamelCase ) return output_string def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [] __lowercase = key - 1 if key <= 0: raise ValueError("""Height of grid can't be 0 or negative""" ) if key == 1: return input_string __lowercase = [[] for _ in range(lowerCamelCase )] # generates template for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowercase = 0 for row in temp_grid: # fills in the characters __lowercase = input_string[counter : counter + len(lowerCamelCase )] grid.append(list(lowerCamelCase ) ) counter += len(lowerCamelCase ) __lowercase = """""" # reads as zigzag for position in range(len(lowerCamelCase ) ): __lowercase = position % (lowest * 2) # puts it in bounds __lowercase = min(lowerCamelCase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = {} for key_guess in range(1 , len(lowerCamelCase ) ): # tries every key __lowercase = decrypt(lowerCamelCase , lowerCamelCase ) return results if __name__ == "__main__": import doctest doctest.testmod()
80
1
def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' def count_of_possible_combinations(lowerCamelCase ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' def count_of_possible_combinations_with_dp_array( lowerCamelCase , lowerCamelCase ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] __lowercase = sum( count_of_possible_combinations_with_dp_array(target - item , lowerCamelCase ) for item in array ) __lowercase = answer return answer __lowercase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(lowerCamelCase , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = [0] * (target + 1) __lowercase = 1 for i in range(1 , target + 1 ): for j in range(lowerCamelCase ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Union[str, Any] = 3 __UpperCamelCase : Any = 5 __UpperCamelCase : Any = [1, 2, 5] print(combination_sum_iv(n, array, target))
80
def snake_case ( lowerCamelCase = 2_000_000 ): '''simple docstring''' __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCamelCase ): __lowercase = 1 __lowercase = 0 for i in range(lowerCamelCase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'''{solution() = }''')
80
1
from __future__ import annotations def snake_case ( lowerCamelCase ): '''simple docstring''' if not nums: raise ValueError("""List is empty""" ) return sum(lowerCamelCase ) / len(lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
80
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int]=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : Optional[Any]=2 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : List[str]=16 , _lowerCAmelCase : List[str]=[1, 2, 1] , _lowerCAmelCase : Dict=[2, 2, 4] , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Optional[Any]=2.0 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : Optional[int]=0.0 , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Optional[int]=0.1 , _lowerCAmelCase : Tuple="gelu" , _lowerCAmelCase : int=False , _lowerCAmelCase : Dict=True , _lowerCAmelCase : List[str]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Tuple=8 , _lowerCAmelCase : List[Any]=["stage1", "stage2", "stage3"] , _lowerCAmelCase : Union[str, Any]=[1, 2, 3] , ) -> int: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def _a ( self : List[Any] ) -> int: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : Dict ) -> Dict: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _a ( self : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : int ) -> Dict: """simple docstring""" __lowercase = MaskFormerSwinModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowerCAmelCase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=_lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __snake_case :Optional[int] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __snake_case :Optional[int] = False __snake_case :Any = False __snake_case :List[str] = False __snake_case :Tuple = False __snake_case :Optional[int] = False def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" pass def _a ( self : Dict ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : List[Any] ) -> Any: """simple docstring""" return def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def _a ( self : Tuple ) -> str: """simple docstring""" pass def _a ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def _a ( self : Optional[int] ) -> int: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def _a ( self : Any ) -> Any: """simple docstring""" pass def _a ( self : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Dict ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def _a ( self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def _a ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowerCAmelCase : Optional[int] ): __lowercase = 0 return t def check_equivalence(_lowerCAmelCase : int , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]={} ): with torch.no_grad(): __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = model(**_lowerCAmelCase , return_dict=_lowerCAmelCase , **_lowerCAmelCase ).to_tuple() def recursive_check(_lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ): if isinstance(_lowerCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCAmelCase , _lowerCAmelCase ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowerCAmelCase , _lowerCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowerCAmelCase ) , set_nan_tensor_to_zero(_lowerCAmelCase ) , atol=1e-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}. Dict has' F' `nan`: {torch.isnan(_lowerCAmelCase ).any()} and `inf`: {torch.isinf(_lowerCAmelCase )}.' ) , ) recursive_check(_lowerCAmelCase , _lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) __lowercase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) check_equivalence(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , {"""output_hidden_states""": True} ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Optional[Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () __snake_case :Dict = MaskFormerSwinConfig def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(_lowerCAmelCase ) backbone.to(_lowerCAmelCase ) backbone.eval() __lowercase = backbone(**_lowerCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowerCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**_lowerCAmelCase , output_hidden_states=_lowerCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**_lowerCAmelCase , output_attentions=_lowerCAmelCase ) self.assertIsNotNone(outputs.attentions )
80
1
def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = len(lowerCamelCase ) + 1 __lowercase = len(lowerCamelCase ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __lowercase = [[0 for i in range(lowerCamelCase )] for j in range(lowerCamelCase )] # since string of zero length match pattern of zero length __lowercase = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , lowerCamelCase ): __lowercase = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , lowerCamelCase ): __lowercase = dp[0][j - 2] if pattern[j - 1] == """*""" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , lowerCamelCase ): for j in range(1 , lowerCamelCase ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __lowercase = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __lowercase = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __lowercase = dp[i - 1][j] else: __lowercase = 0 else: __lowercase = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") __UpperCamelCase : Any = """aab""" __UpperCamelCase : Optional[Any] = """c*a*b""" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F'''{input_string} matches the given pattern {pattern}''') else: print(F'''{input_string} does not match with the given pattern {pattern}''')
80
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.nn.Linear(10 , 10 ) __lowercase = torch.optim.SGD(model.parameters() , 0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F'Accelerated optimizer pickling failed with {e}' ) AcceleratorState._reset_state()
80
1
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : str=13 , _lowerCAmelCase : List[Any]=64 , _lowerCAmelCase : int=2 , _lowerCAmelCase : Any=3 , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Optional[Any]=32 , _lowerCAmelCase : int=5 , _lowerCAmelCase : List[Any]=4 , _lowerCAmelCase : int=37 , _lowerCAmelCase : Any="gelu" , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : Optional[Any]=0.1 , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : Union[str, Any]=0.02 , _lowerCAmelCase : Tuple=[1, 16, 4, 4] , _lowerCAmelCase : str=None , ) -> str: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size __lowercase = (self.image_size // 32) ** 2 __lowercase = num_patches + 1 def _a ( self : int ) -> List[Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : str ) -> List[str]: """simple docstring""" __lowercase = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, """hidden_sizes""": [4, 8, 16, 32], """num_groups""": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=_lowerCAmelCase , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = ViTHybridModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] ) -> Any: """simple docstring""" __lowercase = self.type_sequence_label_size __lowercase = ViTHybridForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Union[str, Any] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __snake_case :int = ( {'feature-extraction': ViTHybridModel, 'image-classification': ViTHybridForImageClassification} if is_torch_available() else {} ) __snake_case :Optional[Any] = False __snake_case :Union[str, Any] = False __snake_case :List[Any] = False def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = ViTHybridModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : int ) -> Dict: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def _a ( self : int ) -> Optional[Any]: """simple docstring""" pass def _a ( self : str ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def _a ( self : Dict ) -> List[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Tuple ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = _config_zero_init(_lowerCAmelCase ) for model_class in self.all_model_classes: __lowercase = model_class(config=_lowerCAmelCase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": __lowercase = [F'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def _a ( self : Optional[Any] ) -> Dict: """simple docstring""" for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ViTHybridModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Optional[int] ) -> str: """simple docstring""" return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _a ( self : List[Any] ) -> Any: """simple docstring""" __lowercase = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( _lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-1.9_090, -0.4_993, -0.2_389] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @slow @require_accelerate def _a ( self : Optional[Any] ) -> str: """simple docstring""" __lowercase = ViTHybridImageProcessor.from_pretrained("""google/vit-hybrid-base-bit-384""" ) __lowercase = ViTHybridForImageClassification.from_pretrained("""google/vit-hybrid-base-bit-384""" , device_map="""auto""" ) __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ) __lowercase = model(**_lowerCAmelCase ) __lowercase = outputs.logits # model predicts one of the 1000 ImageNet classes __lowercase = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , """tabby, tabby cat""" )
80
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __UpperCamelCase : Dict = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __UpperCamelCase : int = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __UpperCamelCase : List[str] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
from math import isqrt def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowerCamelCase , lowerCamelCase ): __lowercase = False return [i for i in range(2 , lowerCamelCase ) if is_prime[i]] def snake_case ( lowerCamelCase = 10**8 ): '''simple docstring''' __lowercase = calculate_prime_numbers(max_number // 2 ) __lowercase = 0 __lowercase = 0 __lowercase = len(lowerCamelCase ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(F'''{solution() = }''')
80
import os from collections.abc import Iterator def snake_case ( lowerCamelCase = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(lowerCamelCase ): __lowercase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(lowerCamelCase , lowerCamelCase ).lstrip("""./""" ) def snake_case ( lowerCamelCase ): '''simple docstring''' return F'{i * " "}*' if i else "\n##" def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' __lowercase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def snake_case ( lowerCamelCase = "." ): '''simple docstring''' __lowercase = """""" for filepath in sorted(good_file_paths(lowerCamelCase ) ): __lowercase , __lowercase = os.path.split(lowerCamelCase ) if filepath != old_path: __lowercase = print_path(lowerCamelCase , lowerCamelCase ) __lowercase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowercase = F'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowercase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(F'{md_prefix(lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
80
1
from __future__ import annotations from typing import Any class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 0 ) -> None: """simple docstring""" __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(_lowerCAmelCase )] for r in range(_lowerCAmelCase )] def __str__( self : List[Any] ) -> str: """simple docstring""" __lowercase = F'Matrix consist of {self.row} rows and {self.column} columns\n' # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(_lowerCAmelCase , len(str(_lowerCAmelCase ) ) ) __lowercase = F'%{max_element_length}s' # Make string and return def single_line(_lowerCAmelCase : list[float] ) -> str: nonlocal string_format_identifier __lowercase = """[""" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(_lowerCAmelCase ) for row_vector in self.array ) return s def __repr__( self : int ) -> str: """simple docstring""" return str(self ) def _a ( self : Dict , _lowerCAmelCase : tuple[int, int] ) -> bool: """simple docstring""" if not (isinstance(_lowerCAmelCase , (list, tuple) ) and len(_lowerCAmelCase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : List[str] , _lowerCAmelCase : tuple[int, int] ) -> Any: """simple docstring""" assert self.validate_indicies(_lowerCAmelCase ) return self.array[loc[0]][loc[1]] def __setitem__( self : List[Any] , _lowerCAmelCase : tuple[int, int] , _lowerCAmelCase : float ) -> None: """simple docstring""" assert self.validate_indicies(_lowerCAmelCase ) __lowercase = value def __add__( self : Any , _lowerCAmelCase : Matrix ) -> Matrix: """simple docstring""" assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : str ) -> Matrix: """simple docstring""" __lowercase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : Tuple , _lowerCAmelCase : Matrix ) -> Matrix: """simple docstring""" return self + (-another) def __mul__( self : Dict , _lowerCAmelCase : int | float | Matrix ) -> Matrix: """simple docstring""" if isinstance(_lowerCAmelCase , (int, float) ): # Scalar multiplication __lowercase = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F'Unsupported type given for another ({type(_lowerCAmelCase )})' raise TypeError(_lowerCAmelCase ) def _a ( self : Any ) -> Matrix: """simple docstring""" __lowercase = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def _a ( self : Dict , _lowerCAmelCase : Matrix , _lowerCAmelCase : Matrix ) -> Any: """simple docstring""" assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def snake_case ( ): '''simple docstring''' __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F'a^(-1) is {ainv}' ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F'u is {u}' ) print(F'v is {v}' ) print(F'uv^T is {u * v.transpose()}' ) # Sherman Morrison print(F'(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase , lowerCamelCase )}' ) def snake_case ( ): '''simple docstring''' import doctest doctest.testmod() testa()
80
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Optional[Any] = logging.get_logger(__name__) __UpperCamelCase : int = { """caidas/swin2sr-classicalsr-x2-64""": ( """https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json""" ), } class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = 'swin2sr' __snake_case :Any = { 'hidden_size': 'embed_dim', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : Union[str, Any] , _lowerCAmelCase : List[str]=64 , _lowerCAmelCase : Any=1 , _lowerCAmelCase : int=3 , _lowerCAmelCase : str=180 , _lowerCAmelCase : Dict=[6, 6, 6, 6, 6, 6] , _lowerCAmelCase : Optional[Any]=[6, 6, 6, 6, 6, 6] , _lowerCAmelCase : Any=8 , _lowerCAmelCase : Any=2.0 , _lowerCAmelCase : str=True , _lowerCAmelCase : Dict=0.0 , _lowerCAmelCase : Dict=0.0 , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : int="gelu" , _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : Optional[Any]=0.02 , _lowerCAmelCase : Union[str, Any]=1e-5 , _lowerCAmelCase : int=2 , _lowerCAmelCase : Dict=1.0 , _lowerCAmelCase : List[str]="1conv" , _lowerCAmelCase : Dict="pixelshuffle" , **_lowerCAmelCase : Optional[Any] , ) -> str: """simple docstring""" super().__init__(**_lowerCAmelCase ) __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = len(_lowerCAmelCase ) __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = upscale __lowercase = img_range __lowercase = resi_connection __lowercase = upsampler
80
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def snake_case ( ): '''simple docstring''' __lowercase = [randint(-1_000 , 1_000 ) for i in range(10 )] __lowercase = randint(-5_000 , 5_000 ) return (arr, r) __UpperCamelCase : Any = make_dataset() def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' for triplet in permutations(lowerCamelCase , 3 ): if sum(lowerCamelCase ) == target: return tuple(sorted(lowerCamelCase ) ) return (0, 0, 0) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' arr.sort() __lowercase = len(lowerCamelCase ) for i in range(n - 1 ): __lowercase , __lowercase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def snake_case ( ): '''simple docstring''' __lowercase = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ __lowercase = """ triplet_sum1(*dataset) """ __lowercase = """ triplet_sum2(*dataset) """ __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) __lowercase = repeat(setup=lowerCamelCase , stmt=lowerCamelCase , repeat=5 , number=10_000 ) return (min(lowerCamelCase ), min(lowerCamelCase )) if __name__ == "__main__": from doctest import testmod testmod() __UpperCamelCase : Tuple = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
80
1
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase : str = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right __UpperCamelCase : List[Any] = 250004 __UpperCamelCase : Optional[int] = 250020 @require_sentencepiece @require_tokenizers class __UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[int] = MBartTokenizer __snake_case :int = MBartTokenizerFast __snake_case :Union[str, Any] = True __snake_case :List[str] = True def _a ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowercase = MBartTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _a ( self : Tuple ) -> Any: """simple docstring""" __lowercase = MBartTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) __lowercase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(_lowerCAmelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowercase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) __lowercase = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __lowercase = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def _a ( self : Any ) -> Union[str, Any]: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __lowercase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-mbart""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): __lowercase = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) __lowercase = tempfile.mkdtemp() __lowercase = tokenizer_r.save_pretrained(_lowerCAmelCase ) __lowercase = tokenizer_p.save_pretrained(_lowerCAmelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) __lowercase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(_lowerCAmelCase , _lowerCAmelCase ) # Checks everything loads correctly in the same way __lowercase = tokenizer_r.from_pretrained(_lowerCAmelCase ) __lowercase = tokenizer_p.from_pretrained(_lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_lowerCAmelCase ) # Save tokenizer rust, legacy_format=True __lowercase = tempfile.mkdtemp() __lowercase = tokenizer_r.save_pretrained(_lowerCAmelCase , legacy_format=_lowerCAmelCase ) __lowercase = tokenizer_p.save_pretrained(_lowerCAmelCase ) # Checks it save with the same files self.assertSequenceEqual(_lowerCAmelCase , _lowerCAmelCase ) # Checks everything loads correctly in the same way __lowercase = tokenizer_r.from_pretrained(_lowerCAmelCase ) __lowercase = tokenizer_p.from_pretrained(_lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase ) ) shutil.rmtree(_lowerCAmelCase ) # Save tokenizer rust, legacy_format=False __lowercase = tempfile.mkdtemp() __lowercase = tokenizer_r.save_pretrained(_lowerCAmelCase , legacy_format=_lowerCAmelCase ) __lowercase = tokenizer_p.save_pretrained(_lowerCAmelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __lowercase = tokenizer_r.from_pretrained(_lowerCAmelCase ) __lowercase = tokenizer_p.from_pretrained(_lowerCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowerCAmelCase , _lowerCAmelCase ) ) shutil.rmtree(_lowerCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): __snake_case :Tuple = 'facebook/mbart-large-en-ro' __snake_case :List[Any] = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] __snake_case :str = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] __snake_case :Tuple = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def _a ( cls : Tuple ) -> Optional[Any]: """simple docstring""" __lowercase = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en_XX""" , tgt_lang="""ro_RO""" ) __lowercase = 1 return cls def _a ( self : Any ) -> str: """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ar_AR"""] , 25_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""en_EN"""] , 25_0004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ro_RO"""] , 25_0020 ) def _a ( self : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) def _a ( self : List[str] ) -> int: """simple docstring""" self.assertIn(_lowerCAmelCase , self.tokenizer.all_special_ids ) __lowercase = [RO_CODE, 884, 9019, 96, 9, 916, 8_6792, 36, 1_8743, 1_5596, 5, 2] __lowercase = self.tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) __lowercase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertNotIn(self.tokenizer.eos_token , _lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , _lowerCAmelCase ) __lowercase = 10 __lowercase = self.tokenizer(_lowerCAmelCase , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , _lowerCAmelCase ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) def _a ( self : Dict ) -> str: """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [25_0026, 25_0001] ) def _a ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = tempfile.mkdtemp() __lowercase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowerCAmelCase ) __lowercase = MBartTokenizer.from_pretrained(_lowerCAmelCase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _lowerCAmelCase ) @require_torch def _a ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , return_tensors="""pt""" ) __lowercase = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) __lowercase = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __lowercase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def _a ( self : Optional[Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer(self.src_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=3 , return_tensors="""pt""" ) __lowercase = self.tokenizer( text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=10 , return_tensors="""pt""" ) __lowercase = targets["""input_ids"""] __lowercase = shift_tokens_right(_lowerCAmelCase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _a ( self : Dict ) -> int: """simple docstring""" __lowercase = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""en_XX""" , tgt_lang="""ar_AR""" ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , { # A, test, EOS, en_XX """input_ids""": [[62, 3034, 2, 25_0004]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 25_0001, } , )
80
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever __UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None ) -> int: """simple docstring""" super().__init__( _lowerCAmelCase , question_encoder_tokenizer=_lowerCAmelCase , generator_tokenizer=_lowerCAmelCase , index=_lowerCAmelCase , init_retrieval=_lowerCAmelCase , ) __lowercase = None def _a ( self : int , _lowerCAmelCase : int ) -> Any: """simple docstring""" logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually __lowercase = self._infer_socket_ifname() # avoid clash with the NCCL port __lowercase = str(distributed_port + 1 ) __lowercase = dist.new_group(ranks=_lowerCAmelCase , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def _a ( self : Tuple ) -> List[str]: """simple docstring""" return dist.get_rank(group=self.process_group ) == 0 def _a ( self : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=torch.floataa ) -> Tuple: """simple docstring""" __lowercase = torch.empty(_lowerCAmelCase , dtype=_lowerCAmelCase ) dist.scatter(_lowerCAmelCase , src=0 , scatter_list=_lowerCAmelCase , group=self.process_group ) return target_tensor def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __lowercase = next((addr for addr in addrs if addr.startswith("""e""" )) , _lowerCAmelCase ) return ifname def _a ( self : str , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: """simple docstring""" if not dist.is_initialized(): __lowercase , __lowercase = self._main_retrieve(_lowerCAmelCase , _lowerCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowerCAmelCase ) # distributed training __lowercase = dist.get_world_size(group=self.process_group ) # gather logic __lowercase = None if self._is_main(): __lowercase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowerCAmelCase )] dist.gather(torch.tensor(_lowerCAmelCase ) , dst=0 , gather_list=_lowerCAmelCase , group=self.process_group ) # scatter logic __lowercase = question_hidden_states.shape[0] __lowercase = [] __lowercase = [] if self._is_main(): assert len(_lowerCAmelCase ) == world_size __lowercase , __lowercase = self._main_retrieve(torch.cat(_lowerCAmelCase ).numpy() , _lowerCAmelCase ) __lowercase , __lowercase = torch.tensor(_lowerCAmelCase ), torch.tensor(_lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._chunk_tensor(_lowerCAmelCase , _lowerCAmelCase ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) __lowercase = self._scattered(_lowerCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowerCAmelCase )
80
1
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def snake_case ( *lowerCamelCase ): '''simple docstring''' with open(lowerCamelCase , """r""" ) as fh: fcntl.flock(lowerCamelCase , fcntl.LOCK_EX ) try: print(*lowerCamelCase ) finally: fcntl.flock(lowerCamelCase , fcntl.LOCK_UN ) __UpperCamelCase : List[Any] = int(os.environ["""LOCAL_RANK"""]) torch.cuda.set_device(local_rank) __UpperCamelCase : str = torch.device("""cuda""", local_rank) __UpperCamelCase : str = socket.gethostname() __UpperCamelCase : Optional[int] = F'''[{hostname}-{local_rank}]''' try: # test distributed dist.init_process_group("""nccl""") dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __UpperCamelCase : Dict = dist.get_rank() __UpperCamelCase : Optional[int] = dist.get_world_size() printflock(F'''{gpu} is OK (global rank: {rank}/{world_size})''') dist.barrier() if rank == 0: printflock(F'''pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}''') except Exception: printflock(F'''{gpu} is broken''') raise
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ): __snake_case :List[Any] = 1 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1000 , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Optional[int]: """simple docstring""" self.set_timesteps(_lowerCAmelCase ) # standard deviation of the initial noise distribution __lowercase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowercase = 4 # running values __lowercase = [] def _a ( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" __lowercase = num_inference_steps __lowercase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowercase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowercase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowercase = torch.sin(steps * math.pi / 2 ) ** 2 __lowercase = (1.0 - self.betas**2) ** 0.5 __lowercase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowercase = timesteps.to(_lowerCAmelCase ) __lowercase = [] def _a ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowercase = (self.timesteps == timestep).nonzero().item() __lowercase = timestep_index + 1 __lowercase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(_lowerCAmelCase ) if len(self.ets ) == 1: __lowercase = self.ets[-1] elif len(self.ets ) == 2: __lowercase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowercase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowercase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowercase = self._get_prev_sample(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _a ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , *_lowerCAmelCase : Optional[Any] , **_lowerCAmelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _a ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Dict: """simple docstring""" __lowercase = self.alphas[timestep_index] __lowercase = self.betas[timestep_index] __lowercase = self.alphas[prev_timestep_index] __lowercase = self.betas[prev_timestep_index] __lowercase = (sample - sigma * ets) / max(_lowerCAmelCase , 1e-8 ) __lowercase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Optional[Any] ) -> Dict: """simple docstring""" return self.config.num_train_timesteps
80
1
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __UpperCamelCase : Tuple = TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] , _lowerCAmelCase : T ) -> List[str]: """simple docstring""" __lowercase = data __lowercase = None def __str__( self : List[str] ) -> str: """simple docstring""" return F'{self.data}' class __UpperCamelCase ( Generic[T] ): def __init__( self : Optional[Any] ) -> None: """simple docstring""" __lowercase = None def __iter__( self : int ) -> Iterator[T]: """simple docstring""" __lowercase = self.top while node: yield node.data __lowercase = node.next def __str__( self : List[str] ) -> str: """simple docstring""" return "->".join([str(_lowerCAmelCase ) for item in self] ) def __len__( self : Any ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _a ( self : str ) -> bool: """simple docstring""" return self.top is None def _a ( self : List[str] , _lowerCAmelCase : T ) -> None: """simple docstring""" __lowercase = Node(_lowerCAmelCase ) if not self.is_empty(): __lowercase = self.top __lowercase = node def _a ( self : Union[str, Any] ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , _lowerCAmelCase ) __lowercase = self.top __lowercase = self.top.next return pop_node.data def _a ( self : int ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _a ( self : int ) -> None: """simple docstring""" __lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
80
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class __UpperCamelCase : __snake_case :Optional[Any] = MBartConfig __snake_case :str = {} __snake_case :str = 'gelu' def __init__( self : Dict , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str]=13 , _lowerCAmelCase : Union[str, Any]=7 , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : List[Any]=False , _lowerCAmelCase : Tuple=99 , _lowerCAmelCase : Dict=32 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Optional[int]=4 , _lowerCAmelCase : str=37 , _lowerCAmelCase : Dict=0.1 , _lowerCAmelCase : Tuple=0.1 , _lowerCAmelCase : Optional[Any]=20 , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : int=1 , _lowerCAmelCase : List[str]=0 , ) -> Tuple: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = eos_token_id __lowercase = pad_token_id __lowercase = bos_token_id def _a ( self : int ) -> Dict: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowercase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowercase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __lowercase = prepare_mbart_inputs_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return config, inputs_dict def _a ( self : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = TFMBartModel(config=_lowerCAmelCase ).get_decoder() __lowercase = inputs_dict["""input_ids"""] __lowercase = input_ids[:1, :] __lowercase = inputs_dict["""attention_mask"""][:1, :] __lowercase = inputs_dict["""head_mask"""] __lowercase = 1 # first forward pass __lowercase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , head_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase ) __lowercase , __lowercase = outputs.to_tuple() __lowercase = past_key_values[1] def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): '''simple docstring''' if attention_mask is None: __lowercase = tf.cast(tf.math.not_equal(lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __lowercase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __lowercase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __lowercase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __lowercase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :List[Any] = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () __snake_case :Optional[int] = (TFMBartForConditionalGeneration,) if is_tf_available() else () __snake_case :Any = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) __snake_case :str = True __snake_case :List[str] = False __snake_case :Tuple = False def _a ( self : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] ) -> Tuple: """simple docstring""" if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = TFMBartModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase ) def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def _a ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowerCAmelCase ) @require_sentencepiece @require_tokenizers @require_tf class __UpperCamelCase ( unittest.TestCase ): __snake_case :Union[str, Any] = [ ' UN Chief Says There Is No Military Solution in Syria', ] __snake_case :Optional[Any] = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] __snake_case :List[Any] = 'facebook/mbart-large-en-ro' @cached_property def _a ( self : List[Any] ) -> Any: """simple docstring""" return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def _a ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def _a ( self : str , **_lowerCAmelCase : Tuple ) -> int: """simple docstring""" __lowercase = self.translate_src_text(**_lowerCAmelCase ) self.assertListEqual(self.expected_text , _lowerCAmelCase ) def _a ( self : Dict , **_lowerCAmelCase : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer(self.src_text , **_lowerCAmelCase , return_tensors="""tf""" ) __lowercase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) __lowercase = self.tokenizer.batch_decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) return generated_words @slow def _a ( self : List[str] ) -> List[Any]: """simple docstring""" self._assert_generated_batch_equal_expected()
80
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device __UpperCamelCase : Union[str, Any] = False class __UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Any ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_lowerCAmelCase ) __lowercase = VersatileDiffusionPipeline.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt="""first prompt""" , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _a ( self : Any ) -> Dict: """simple docstring""" __lowercase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __lowercase = """cyberpunk 2077""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowercase = torch.manual_seed(0 ) __lowercase = pipe.dual_guided( prompt=_lowerCAmelCase , image=_lowerCAmelCase , text_to_image_strength=0.75 , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = """A painting of a squirrel eating a burger """ __lowercase = torch.manual_seed(0 ) __lowercase = pipe.text_to_image( prompt=_lowerCAmelCase , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 __lowercase = pipe.image_variation(_lowerCAmelCase , generator=_lowerCAmelCase , output_type="""numpy""" ).images __lowercase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowercase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
80
1
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __UpperCamelCase : Optional[int] = logging.get_logger(__name__) class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :Dict = ['pixel_values'] def __init__( self : int , _lowerCAmelCase : bool = True , _lowerCAmelCase : Union[int, float] = 1 / 255 , _lowerCAmelCase : bool = True , _lowerCAmelCase : int = 8 , **_lowerCAmelCase : str , ) -> None: """simple docstring""" super().__init__(**_lowerCAmelCase ) __lowercase = do_rescale __lowercase = rescale_factor __lowercase = do_pad __lowercase = pad_size def _a ( self : Any , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : float , _lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None , **_lowerCAmelCase : Any ) -> np.ndarray: """simple docstring""" return rescale(_lowerCAmelCase , scale=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Union[str, ChannelDimension]] = None ) -> str: """simple docstring""" __lowercase , __lowercase = get_image_size(_lowerCAmelCase ) __lowercase = (old_height // size + 1) * size - old_height __lowercase = (old_width // size + 1) * size - old_width return pad(_lowerCAmelCase , ((0, pad_height), (0, pad_width)) , mode="""symmetric""" , data_format=_lowerCAmelCase ) def _a ( self : List[Any] , _lowerCAmelCase : ImageInput , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[float] = None , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : Optional[Union[str, TensorType]] = None , _lowerCAmelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **_lowerCAmelCase : Any , ) -> Any: """simple docstring""" __lowercase = do_rescale if do_rescale is not None else self.do_rescale __lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase = do_pad if do_pad is not None else self.do_pad __lowercase = pad_size if pad_size is not None else self.pad_size __lowercase = make_list_of_images(_lowerCAmelCase ) if not valid_images(_lowerCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. __lowercase = [to_numpy_array(_lowerCAmelCase ) for image in images] if do_rescale: __lowercase = [self.rescale(image=_lowerCAmelCase , scale=_lowerCAmelCase ) for image in images] if do_pad: __lowercase = [self.pad(_lowerCAmelCase , size=_lowerCAmelCase ) for image in images] __lowercase = [to_channel_dimension_format(_lowerCAmelCase , _lowerCAmelCase ) for image in images] __lowercase = {"""pixel_values""": images} return BatchFeature(data=_lowerCAmelCase , tensor_type=_lowerCAmelCase )
80
from __future__ import annotations from collections.abc import MutableSequence class __UpperCamelCase : def __init__( self : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : MutableSequence[float] ) -> None: """simple docstring""" if len(_lowerCAmelCase ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) __lowercase = list(_lowerCAmelCase ) __lowercase = degree def __add__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: __lowercase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , _lowerCAmelCase ) else: __lowercase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , _lowerCAmelCase ) def __sub__( self : int , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , _lowerCAmelCase : Polynomial ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , _lowerCAmelCase ) def _a ( self : Optional[int] , _lowerCAmelCase : int | float ) -> int | float: """simple docstring""" __lowercase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" __lowercase = """""" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(_lowerCAmelCase ) return polynomial def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return self.__str__() def _a ( self : List[str] ) -> Polynomial: """simple docstring""" __lowercase = [0] * self.degree for i in range(self.degree ): __lowercase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , _lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : int | float = 0 ) -> Polynomial: """simple docstring""" __lowercase = [0] * (self.degree + 2) __lowercase = constant for i in range(self.degree + 1 ): __lowercase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , _lowerCAmelCase ) def __eq__( self : List[str] , _lowerCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Dict , _lowerCAmelCase : object ) -> bool: """simple docstring""" return not self.__eq__(_lowerCAmelCase )
80
1
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Dict = StableUnCLIPPipeline __snake_case :List[Any] = TEXT_TO_IMAGE_PARAMS __snake_case :Dict = TEXT_TO_IMAGE_BATCH_PARAMS __snake_case :Union[str, Any] = TEXT_TO_IMAGE_IMAGE_PARAMS __snake_case :Any = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false __snake_case :Tuple = False def _a ( self : str ) -> Any: """simple docstring""" __lowercase = 32 __lowercase = embedder_hidden_size # prior components torch.manual_seed(0 ) __lowercase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) __lowercase = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowerCAmelCase , projection_dim=_lowerCAmelCase , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) __lowercase = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=_lowerCAmelCase , num_layers=1 , ) torch.manual_seed(0 ) __lowercase = DDPMScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=_lowerCAmelCase , clip_sample_range=5.0 , beta_schedule="""squaredcos_cap_v2""" , ) # regular denoising components torch.manual_seed(0 ) __lowercase = StableUnCLIPImageNormalizer(embedding_dim=_lowerCAmelCase ) __lowercase = DDPMScheduler(beta_schedule="""squaredcos_cap_v2""" ) torch.manual_seed(0 ) __lowercase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) __lowercase = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowerCAmelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="""projection""" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_lowerCAmelCase , layers_per_block=1 , upcast_attention=_lowerCAmelCase , use_linear_projection=_lowerCAmelCase , ) torch.manual_seed(0 ) __lowercase = DDIMScheduler( beta_schedule="""scaled_linear""" , beta_start=0.00_085 , beta_end=0.012 , prediction_type="""v_prediction""" , set_alpha_to_one=_lowerCAmelCase , steps_offset=1 , ) torch.manual_seed(0 ) __lowercase = AutoencoderKL() __lowercase = { # prior components """prior_tokenizer""": prior_tokenizer, """prior_text_encoder""": prior_text_encoder, """prior""": prior, """prior_scheduler""": prior_scheduler, # image noising components """image_normalizer""": image_normalizer, """image_noising_scheduler""": image_noising_scheduler, # regular denoising components """tokenizer""": tokenizer, """text_encoder""": text_encoder, """unet""": unet, """scheduler""": scheduler, """vae""": vae, } return components def _a ( self : int , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any]=0 ) -> List[Any]: """simple docstring""" if str(_lowerCAmelCase ).startswith("""mps""" ): __lowercase = torch.manual_seed(_lowerCAmelCase ) else: __lowercase = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """prior_num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase = torch_device == """cpu""" self._test_attention_slicing_forward_pass(test_max_difference=_lowerCAmelCase ) def _a ( self : Optional[Any] ) -> str: """simple docstring""" __lowercase = torch_device in ["""cpu""", """mps"""] self._test_inference_batch_single_identical(test_max_difference=_lowerCAmelCase ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def _a ( self : Dict ) -> Dict: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : Tuple ) -> Any: """simple docstring""" __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy""" ) __lowercase = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __lowercase = torch.Generator(device="""cpu""" ).manual_seed(0 ) __lowercase = pipe("""anime turle""" , generator=_lowerCAmelCase , output_type="""np""" ) __lowercase = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase ) def _a ( self : List[str] ) -> int: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowercase = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) __lowercase = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __lowercase = pipe( """anime turtle""" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="""np""" , ) __lowercase = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
80
def snake_case ( lowerCamelCase ): '''simple docstring''' if collection == []: return [] # get some information about the collection __lowercase = len(lowerCamelCase ) __lowercase = max(lowerCamelCase ) __lowercase = min(lowerCamelCase ) # create the counting array __lowercase = coll_max + 1 - coll_min __lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCamelCase ): __lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection __lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCamelCase ) ): __lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def snake_case ( lowerCamelCase ): '''simple docstring''' return "".join([chr(lowerCamelCase ) for i in counting_sort([ord(lowerCamelCase ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" __UpperCamelCase : str = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : Union[str, Any] = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
80
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=_lowerCAmelCase ) class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :str = field(default='audio-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) __snake_case :ClassVar[Features] = Features({'audio': Audio()} ) __snake_case :ClassVar[Features] = Features({'labels': ClassLabel} ) __snake_case :str = "audio" __snake_case :str = "labels" def _a ( self : Any , _lowerCAmelCase : Dict ) -> Union[str, Any]: """simple docstring""" if self.label_column not in features: raise ValueError(F'Column {self.label_column} is not present in features.' ) if not isinstance(features[self.label_column] , _lowerCAmelCase ): raise ValueError(F'Column {self.label_column} is not a ClassLabel.' ) __lowercase = copy.deepcopy(self ) __lowercase = self.label_schema.copy() __lowercase = features[self.label_column] __lowercase = label_schema return task_template @property def _a ( self : Dict ) -> Dict[str, str]: """simple docstring""" return { self.audio_column: "audio", self.label_column: "labels", }
80
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase : def __init__( self : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int=13 , _lowerCAmelCase : Any=32 , _lowerCAmelCase : str=3 , _lowerCAmelCase : int=4 , _lowerCAmelCase : Optional[int]=[10, 20, 30, 40] , _lowerCAmelCase : Optional[Any]=[2, 2, 3, 2] , _lowerCAmelCase : List[str]=True , _lowerCAmelCase : int=True , _lowerCAmelCase : List[str]=37 , _lowerCAmelCase : List[str]="gelu" , _lowerCAmelCase : List[Any]=10 , _lowerCAmelCase : int=0.02 , _lowerCAmelCase : str=["stage2", "stage3", "stage4"] , _lowerCAmelCase : Dict=[2, 3, 4] , _lowerCAmelCase : Tuple=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = num_stages __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_labels __lowercase = initializer_range __lowercase = out_features __lowercase = out_indices __lowercase = scope def _a ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def _a ( self : List[str] ) -> Any: """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _a ( self : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" __lowercase = ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _a ( self : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowercase = None __lowercase = ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowercase = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _a ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): __snake_case :Optional[Any] = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __snake_case :List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) __snake_case :str = True __snake_case :Any = False __snake_case :Any = False __snake_case :Any = False __snake_case :int = False def _a ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = ConvNextModelTester(self ) __lowercase = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def _a ( self : Optional[Any] ) -> int: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : Any ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _a ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _a ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" pass def _a ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(_lowerCAmelCase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _a ( self : Any ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _a ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def _a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" def check_hidden_states_output(_lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): __lowercase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _a ( self : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def _a ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def snake_case ( ): '''simple docstring''' __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def _a ( self : Tuple ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _a ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_lowerCAmelCase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): __lowercase = model(**_lowerCAmelCase ) # verify the logits __lowercase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) __lowercase = torch.tensor([-0.0_260, -0.4_739, 0.1_911] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase , _lowerCAmelCase ): __snake_case :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __snake_case :str = ConvNextConfig __snake_case :Optional[Any] = False def _a ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = ConvNextModelTester(self )
80
1
from collections.abc import Generator from math import sin def snake_case ( lowerCamelCase ): '''simple docstring''' if len(lowerCamelCase ) != 32: raise ValueError("""Input must be of length 32""" ) __lowercase = b"""""" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def snake_case ( lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("""Input must be non-negative""" ) __lowercase = format(lowerCamelCase , """08x""" )[-8:] __lowercase = b"""""" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("""utf-8""" ) return little_endian_hex def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = b"""""" for char in message: bit_string += format(lowerCamelCase , """08b""" ).encode("""utf-8""" ) __lowercase = format(len(lowerCamelCase ) , """064b""" ).encode("""utf-8""" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(lowerCamelCase ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def snake_case ( lowerCamelCase ): '''simple docstring''' if len(lowerCamelCase ) % 512 != 0: raise ValueError("""Input must have length that's a multiple of 512""" ) for pos in range(0 , len(lowerCamelCase ) , 512 ): __lowercase = bit_string[pos : pos + 512] __lowercase = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def snake_case ( lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("""Input must be non-negative""" ) __lowercase = format(lowerCamelCase , """032b""" ) __lowercase = """""" for c in i_str: new_str += "1" if c == "0" else "0" return int(lowerCamelCase , 2 ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' return (a + b) % 2**32 def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("""Input must be non-negative""" ) if shift < 0: raise ValueError("""Shift must be non-negative""" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = preprocess(lowerCamelCase ) __lowercase = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states __lowercase = 0x67452301 __lowercase = 0xefcdab89 __lowercase = 0x98badcfe __lowercase = 0x10325476 __lowercase = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(lowerCamelCase ): __lowercase = aa __lowercase = ba __lowercase = ca __lowercase = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f __lowercase = d ^ (b & (c ^ d)) __lowercase = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f __lowercase = c ^ (d & (b ^ c)) __lowercase = (5 * i + 1) % 16 elif i <= 47: __lowercase = b ^ c ^ d __lowercase = (3 * i + 5) % 16 else: __lowercase = c ^ (b | not_aa(lowerCamelCase )) __lowercase = (7 * i) % 16 __lowercase = (f + a + added_consts[i] + block_words[g]) % 2**32 __lowercase = d __lowercase = c __lowercase = b __lowercase = sum_aa(lowerCamelCase , left_rotate_aa(lowerCamelCase , shift_amounts[i] ) ) # Add hashed chunk to running total __lowercase = sum_aa(lowerCamelCase , lowerCamelCase ) __lowercase = sum_aa(lowerCamelCase , lowerCamelCase ) __lowercase = sum_aa(lowerCamelCase , lowerCamelCase ) __lowercase = sum_aa(lowerCamelCase , lowerCamelCase ) __lowercase = reformat_hex(lowerCamelCase ) + reformat_hex(lowerCamelCase ) + reformat_hex(lowerCamelCase ) + reformat_hex(lowerCamelCase ) return digest if __name__ == "__main__": import doctest doctest.testmod()
80
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
1
from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class __UpperCamelCase : __snake_case :int __snake_case :int class __UpperCamelCase : def __init__( self : Any , _lowerCAmelCase : int ) -> List[str]: """simple docstring""" __lowercase = [[] for _ in range(_lowerCAmelCase )] __lowercase = size def __getitem__( self : Tuple , _lowerCAmelCase : int ) -> Iterator[Edge]: """simple docstring""" return iter(self._graph[vertex] ) @property def _a ( self : str ) -> Dict: """simple docstring""" return self._size def _a ( self : Dict , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> Tuple: """simple docstring""" if weight not in (0, 1): raise ValueError("""Edge weight must be either 0 or 1.""" ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("""Vertex indexes must be in [0; size).""" ) self._graph[from_vertex].append(Edge(_lowerCAmelCase , _lowerCAmelCase ) ) def _a ( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> int | None: """simple docstring""" __lowercase = deque([start_vertex] ) __lowercase = [None] * self.size __lowercase = 0 while queue: __lowercase = queue.popleft() __lowercase = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: __lowercase = current_distance + edge.weight __lowercase = distances[edge.destination_vertex] if ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and new_distance >= dest_vertex_distance ): continue __lowercase = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("""No path from start_vertex to finish_vertex.""" ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
80
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : int = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
80
1
def snake_case ( lowerCamelCase ): '''simple docstring''' for i in range(len(lowerCamelCase ) - 1 , 0 , -1 ): __lowercase = False for j in range(lowerCamelCase , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: __lowercase , __lowercase = unsorted[j - 1], unsorted[j] __lowercase = True for j in range(lowerCamelCase ): if unsorted[j] > unsorted[j + 1]: __lowercase , __lowercase = unsorted[j + 1], unsorted[j] __lowercase = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : List[Any] = input("""Enter numbers separated by a comma:\n""").strip() __UpperCamelCase : int = [int(item) for item in user_input.split(""",""")] print(F'''{cocktail_shaker_sort(unsorted) = }''')
80
from sklearn.metrics import matthews_corrcoef import datasets __UpperCamelCase : Union[str, Any] = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __UpperCamelCase : List[str] = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __UpperCamelCase : Tuple = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def _a ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def _a ( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=None ) -> Optional[Any]: """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase , sample_weight=_lowerCAmelCase ) ), }
80
1
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase : str = logging.get_logger(__name__) __UpperCamelCase : Dict = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __UpperCamelCase : Optional[int] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __UpperCamelCase : Dict = {"""facebook/blenderbot_small-90M""": 512} def snake_case ( lowerCamelCase ): '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowercase = char __lowercase = set(lowerCamelCase ) return pairs class __UpperCamelCase ( _lowerCAmelCase ): __snake_case :List[Any] = VOCAB_FILES_NAMES __snake_case :Tuple = PRETRAINED_VOCAB_FILES_MAP __snake_case :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case :str = ['input_ids', 'attention_mask'] def __init__( self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : str="__start__" , _lowerCAmelCase : int="__end__" , _lowerCAmelCase : Any="__unk__" , _lowerCAmelCase : List[Any]="__null__" , **_lowerCAmelCase : Tuple , ) -> str: """simple docstring""" super().__init__(unk_token=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , encoding="""utf-8""" ) as vocab_handle: __lowercase = json.load(_lowerCAmelCase ) __lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCAmelCase , encoding="""utf-8""" ) as merges_handle: __lowercase = merges_handle.read().split("""\n""" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in merges] __lowercase = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) __lowercase = {} @property def _a ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.encoder ) def _a ( self : Dict ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _a ( self : str , _lowerCAmelCase : str ) -> str: """simple docstring""" if token in self.cache: return self.cache[token] __lowercase = re.sub("""([.,!?()])""" , r""" \1""" , _lowerCAmelCase ) __lowercase = re.sub("""(')""" , r""" \1 """ , _lowerCAmelCase ) __lowercase = re.sub(r"""\s{2,}""" , """ """ , _lowerCAmelCase ) if "\n" in token: __lowercase = token.replace("""\n""" , """ __newln__""" ) __lowercase = token.split(""" """ ) __lowercase = [] for token in tokens: if not len(_lowerCAmelCase ): continue __lowercase = token.lower() __lowercase = tuple(_lowerCAmelCase ) __lowercase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowercase = get_pairs(_lowerCAmelCase ) if not pairs: words.append(_lowerCAmelCase ) continue while True: __lowercase = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowercase , __lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(_lowerCAmelCase ): try: __lowercase = word.index(_lowerCAmelCase , _lowerCAmelCase ) new_word.extend(word[i:j] ) __lowercase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(_lowerCAmelCase ) __lowercase = new_word if len(_lowerCAmelCase ) == 1: break else: __lowercase = get_pairs(_lowerCAmelCase ) __lowercase = """@@ """.join(_lowerCAmelCase ) __lowercase = word[:-4] __lowercase = word words.append(_lowerCAmelCase ) return " ".join(_lowerCAmelCase ) def _a ( self : Optional[Any] , _lowerCAmelCase : str ) -> List[str]: """simple docstring""" __lowercase = [] __lowercase = re.findall(r"""\S+\n?""" , _lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCAmelCase ).split(""" """ ) ) ) return split_tokens def _a ( self : Tuple , _lowerCAmelCase : str ) -> int: """simple docstring""" __lowercase = token.lower() return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _a ( self : Tuple , _lowerCAmelCase : int ) -> str: """simple docstring""" return self.decoder.get(_lowerCAmelCase , self.unk_token ) def _a ( self : Dict , _lowerCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = """ """.join(_lowerCAmelCase ).replace("""@@ """ , """""" ).strip() return out_string def _a ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join( _lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + """\n""" ) __lowercase = 0 with open(_lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowercase = token_index writer.write(""" """.join(_lowerCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
80
1