code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def __lowerCAmelCase (_UpperCamelCase="" ): __lowerCAmelCase : Optional[int] = tempfile.mkdtemp() return os.path.join(_UpperCamelCase , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): __lowerCAmelCase : Dict = torch.rand(12 , dtype=torch.floataa ) - 0.5 __lowerCAmelCase : Optional[Any] = AgentAudio(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , agent_type.to_raw() , atol=1E-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(_SCREAMING_SNAKE_CASE ) ) # Ensure that the file contains the same value as the original tensor __lowerCAmelCase , __lowerCAmelCase : int = sf.read(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , torch.tensor(_SCREAMING_SNAKE_CASE ) , atol=1E-4 ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = torch.rand(12 , dtype=torch.floataa ) - 0.5 __lowerCAmelCase : int = get_new_path(suffix='.wav' ) sf.write(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 1_60_00 ) __lowerCAmelCase : int = AgentAudio(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , agent_type.to_raw() , atol=1E-4 ) ) self.assertEqual(agent_type.to_string() , _SCREAMING_SNAKE_CASE ) @require_vision @require_torch class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = torch.randint(0 , 2_56 , (64, 64, 3) ) __lowerCAmelCase : int = AgentImage(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , agent_type._tensor , atol=1E-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Any = Path(get_tests_dir('fixtures/tests_samples/COCO' ) ) / '000000039769.png' __lowerCAmelCase : Tuple = Image.open(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = AgentImage(_SCREAMING_SNAKE_CASE ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Any = Path(get_tests_dir('fixtures/tests_samples/COCO' ) ) / '000000039769.png' __lowerCAmelCase : Any = Image.open(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = AgentImage(_SCREAMING_SNAKE_CASE ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_SCREAMING_SNAKE_CASE ) ) class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): __lowerCAmelCase : List[str] = 'Hey!' __lowerCAmelCase : Optional[Any] = AgentText(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , agent_type.to_string() ) self.assertEqual(_SCREAMING_SNAKE_CASE , agent_type.to_raw() ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )
86
from __future__ import annotations class lowerCamelCase : """simple docstring""" def __init__( self : Union[str, Any] , __magic_name__ : str , __magic_name__ : str ) -> Dict: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = text, pattern SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = len(__magic_name__ ), len(__magic_name__ ) def __A ( self : List[str] , __magic_name__ : str ) -> int: for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def __A ( self : Dict , __magic_name__ : int ) -> int: for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def __A ( self : Tuple ) -> list[int]: # searches pattern in text and returns index positions SCREAMING_SNAKE_CASE_ = [] for i in range(self.textLen - self.patLen + 1 ): SCREAMING_SNAKE_CASE_ = self.mismatch_in_text(__magic_name__ ) if mismatch_index == -1: positions.append(__magic_name__ ) else: SCREAMING_SNAKE_CASE_ = self.match_in_pattern(self.text[mismatch_index] ) SCREAMING_SNAKE_CASE_ = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions A : Dict = "ABAABA" A : Union[str, Any] = "AB" A : str = BoyerMooreSearch(text, pattern) A : Tuple = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
118
0
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timesteps, smartaa_timesteps, smartaaa_timesteps, smartaaa_timesteps, superaa_timesteps, superaa_timesteps, superaaa_timesteps, ) @dataclass class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Union[List[PIL.Image.Image], np.ndarray] _UpperCamelCase : Optional[List[bool]] _UpperCamelCase : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_if import IFPipeline from .pipeline_if_imgaimg import IFImgaImgPipeline from .pipeline_if_imgaimg_superresolution import IFImgaImgSuperResolutionPipeline from .pipeline_if_inpainting import IFInpaintingPipeline from .pipeline_if_inpainting_superresolution import IFInpaintingSuperResolutionPipeline from .pipeline_if_superresolution import IFSuperResolutionPipeline from .safety_checker import IFSafetyChecker from .watermark import IFWatermarker
126
"""simple docstring""" def SCREAMING_SNAKE_CASE ( _lowerCamelCase : str ,_lowerCamelCase : bool = False ) -> str: if not isinstance(_lowerCamelCase ,_lowerCamelCase ): _lowerCAmelCase : Union[str, Any] = f"Expected string as input, found {type(_lowerCamelCase )}" raise ValueError(_lowerCamelCase ) if not isinstance(_lowerCamelCase ,_lowerCamelCase ): _lowerCAmelCase : Union[str, Any] = f"Expected boolean as use_pascal parameter, found {type(_lowerCamelCase )}" raise ValueError(_lowerCamelCase ) _lowerCAmelCase : str = input_str.split("""_""" ) _lowerCAmelCase : Optional[int] = 0 if use_pascal else 1 _lowerCAmelCase : List[Any] = words[start_index:] _lowerCAmelCase : Union[str, Any] = [word[0].upper() + word[1:] for word in words_to_capitalize] _lowerCAmelCase : List[Any] = """""" if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
126
1
'''simple docstring''' from collections import defaultdict from math import gcd def _UpperCamelCase ( __A = 1500000 ) -> int: '''simple docstring''' UpperCamelCase__ = defaultdict(__A ) UpperCamelCase__ = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue UpperCamelCase__ = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F"""{solution() = }""")
80
'''simple docstring''' def lowerCAmelCase_ ( _lowerCamelCase: int ): __SCREAMING_SNAKE_CASE : str = int(_lowerCamelCase ) if n_element < 1: __SCREAMING_SNAKE_CASE : List[str] = ValueError("""a should be a positive number""" ) raise my_error __SCREAMING_SNAKE_CASE : List[Any] = [1] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[int] = (0, 0, 0) __SCREAMING_SNAKE_CASE : List[str] = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": UpperCamelCase__ : Optional[Any] = input('''Enter the last number (nth term) of the Hamming Number Series: ''') print('''Formula of Hamming Number Series => 2^i * 3^j * 5^k''') UpperCamelCase__ : List[str] = hamming(int(n)) print('''-----------------------------------------------------''') print(f"The list with nth numbers is: {hamming_numbers}") print('''-----------------------------------------------------''')
112
0
'''simple docstring''' import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def a ( __a ) -> Union[str, Any]: '''simple docstring''' return (data["data"], data["target"]) def a ( __a , __a , __a ) -> List[str]: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = XGBRegressor(verbosity=0 , random_state=42 ) xgb.fit(a__ , a__ ) # Predict target for test data UpperCamelCase__ :Tuple = xgb.predict(a__ ) UpperCamelCase__ :List[Any] = predictions.reshape(len(a__ ) , 1 ) return predictions def a ( ) -> int: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = fetch_california_housing() UpperCamelCase__ :Union[str, Any] = data_handling(a__ ) UpperCamelCase__ :List[str] = train_test_split( a__ , a__ , test_size=0.2_5 , random_state=1 ) UpperCamelCase__ :Dict = xgboost(a__ , a__ , a__ ) # Error printing print(f'''Mean Absolute Error : {mean_absolute_error(a__ , a__ )}''' ) print(f'''Mean Square Error : {mean_squared_error(a__ , a__ )}''' ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
369
'''simple docstring''' import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( A__ ): """simple docstring""" _a = (DDIMParallelScheduler,) _a = (('eta', 0.0), ('num_inference_steps', 50)) def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''clip_sample''': True, } config.update(**UpperCamelCase_ ) return config def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :str = self.scheduler_classes[0] UpperCamelCase__ :Optional[Any] = self.get_scheduler_config(**UpperCamelCase_ ) UpperCamelCase__ :Optional[Any] = scheduler_class(**UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ :Optional[int] = 10, 0.0 UpperCamelCase__ :List[str] = self.dummy_model() UpperCamelCase__ :List[str] = self.dummy_sample_deter scheduler.set_timesteps(UpperCamelCase_ ) for t in scheduler.timesteps: UpperCamelCase__ :Optional[int] = model(UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase__ :List[Any] = scheduler.step(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ).prev_sample return sample def lowerCAmelCase__ ( self ): '''simple docstring''' for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=UpperCamelCase_ ) UpperCamelCase__ :List[str] = self.scheduler_classes[0] UpperCamelCase__ :List[str] = self.get_scheduler_config(steps_offset=1 ) UpperCamelCase__ :str = scheduler_class(**UpperCamelCase_ ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1] ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=UpperCamelCase_ , beta_end=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.check_over_configs(thresholding=UpperCamelCase_ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=UpperCamelCase_ , prediction_type=UpperCamelCase_ , sample_max_value=UpperCamelCase_ , ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t in [1, 10, 49]: self.check_over_forward(time_step=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500] ): self.check_over_forward(time_step=UpperCamelCase_ , num_inference_steps=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=UpperCamelCase_ , eta=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = self.scheduler_classes[0] UpperCamelCase__ :List[str] = self.get_scheduler_config() UpperCamelCase__ :int = scheduler_class(**UpperCamelCase_ ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400 ) - 0.14771 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960 ) - 0.32460 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486 ) - 0.00979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998 ) - 0.02 ) ) < 1e-5 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.scheduler_classes[0] UpperCamelCase__ :Dict = self.get_scheduler_config() UpperCamelCase__ :Any = scheduler_class(**UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ :List[str] = 10, 0.0 scheduler.set_timesteps(UpperCamelCase_ ) UpperCamelCase__ :Optional[int] = self.dummy_model() UpperCamelCase__ :List[Any] = self.dummy_sample_deter UpperCamelCase__ :int = self.dummy_sample_deter + 0.1 UpperCamelCase__ :List[str] = self.dummy_sample_deter - 0.1 UpperCamelCase__ :str = samplea.shape[0] UpperCamelCase__ :Optional[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) UpperCamelCase__ :List[Any] = torch.arange(UpperCamelCase_ )[0:3, None].repeat(1 , UpperCamelCase_ ) UpperCamelCase__ :Dict = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) UpperCamelCase__ :Any = scheduler.batch_step_no_noise(UpperCamelCase_ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , UpperCamelCase_ ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :Any = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 1147.7904 ) < 1e-2 assert abs(result_mean.item() - 0.4982 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.full_loop() UpperCamelCase__ :List[str] = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :Any = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 172.0067 ) < 1e-2 assert abs(result_mean.item() - 0.223967 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.full_loop(prediction_type='''v_prediction''' ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :str = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 52.5302 ) < 1e-2 assert abs(result_mean.item() - 0.0684 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.full_loop(set_alpha_to_one=UpperCamelCase_ , beta_start=0.01 ) UpperCamelCase__ :str = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :int = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 149.8295 ) < 1e-2 assert abs(result_mean.item() - 0.1951 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = self.full_loop(set_alpha_to_one=UpperCamelCase_ , beta_start=0.01 ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :List[str] = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 149.0784 ) < 1e-2 assert abs(result_mean.item() - 0.1941 ) < 1e-3
219
0
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Dict: UpperCamelCase__ : Optional[int] = [0 for i in range(len(__lowerCamelCase ) )] # initialize interval's left pointer and right pointer UpperCamelCase__ : Any = 0, 0 for i in range(1 , len(__lowerCamelCase ) ): # case when current index is inside the interval if i <= right_pointer: UpperCamelCase__ : Optional[Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) UpperCamelCase__ : List[Any] = min_edge while go_next(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: UpperCamelCase__ : Dict = i, i + z_result[i] - 1 return z_result def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: return i + z_result[i] < len(__lowerCamelCase ) and s[z_result[i]] == s[i + z_result[i]] def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Any: UpperCamelCase__ : Dict = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string UpperCamelCase__ : Any = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(__lowerCamelCase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
189
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _snake_case : int = logging.get_logger(__name__) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : str = None , lowerCamelCase : uuid.UUID = None , lowerCamelCase : Dict=None , lowerCamelCase : Union[str, Any]=None ) -> int: if not conversation_id: __snake_case : Optional[Any] = uuid.uuida() if past_user_inputs is None: __snake_case : List[Any] = [] if generated_responses is None: __snake_case : Optional[int] = [] __snake_case : uuid.UUID = conversation_id __snake_case : List[str] = past_user_inputs __snake_case : List[str] = generated_responses __snake_case : Optional[str] = text def __eq__( self : int , lowerCamelCase : List[str] ) -> Dict: if not isinstance(lowerCamelCase , lowerCamelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __snake_case ( self : Optional[int] , lowerCamelCase : str , lowerCamelCase : bool = False ) -> Optional[Any]: if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __snake_case : Dict = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __snake_case : List[str] = text def __snake_case ( self : List[str] ) -> Dict: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __snake_case : Optional[Any] = None def __snake_case ( self : List[str] , lowerCamelCase : str ) -> List[Any]: self.generated_responses.append(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Optional[Any] ) -> Dict: __snake_case : Any = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __snake_case : List[Any] = "user" if is_user else "bot" output += F'{name} >> {text} \n' return output @add_end_docstrings( _lowerCAmelCase , R"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : Optional[Any] , **lowerCamelCase : List[str] ) -> Any: super().__init__(*lowerCamelCase , **lowerCamelCase ) if self.tokenizer.pad_token_id is None: __snake_case : Dict = self.tokenizer.eos_token def __snake_case ( self : Dict , lowerCamelCase : List[str]=None , lowerCamelCase : int=None , lowerCamelCase : Optional[Any]=None , **lowerCamelCase : Any ) -> Any: __snake_case : Union[str, Any] = {} __snake_case : Optional[int] = {} __snake_case : Optional[Any] = {} if min_length_for_response is not None: __snake_case : int = min_length_for_response if minimum_tokens is not None: __snake_case : Tuple = minimum_tokens if "max_length" in generate_kwargs: __snake_case : Any = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __snake_case : Any = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowerCamelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , lowerCamelCase : Union[Conversation, List[Conversation]] , lowerCamelCase : Optional[Any]=0 , **lowerCamelCase : Optional[Any] ) -> Union[str, Any]: __snake_case : Optional[Any] = super().__call__(lowerCamelCase , num_workers=lowerCamelCase , **lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) and len(lowerCamelCase ) == 1: return outputs[0] return outputs def __snake_case ( self : Any , lowerCamelCase : Conversation , lowerCamelCase : Any=32 ) -> Dict[str, Any]: if not isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError("ConversationalPipeline, expects Conversation as inputs" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method" ) if hasattr(self.tokenizer , "_build_conversation_input_ids" ): __snake_case : Tuple = self.tokenizer._build_conversation_input_ids(lowerCamelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version __snake_case : Tuple = self._legacy_parse_and_tokenize(lowerCamelCase ) if self.framework == "pt": __snake_case : Union[str, Any] = torch.LongTensor([input_ids] ) elif self.framework == "tf": __snake_case : Union[str, Any] = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def __snake_case ( self : List[str] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict=10 , **lowerCamelCase : Tuple ) -> List[str]: __snake_case : Tuple = generate_kwargs.get("max_length" , self.model.config.max_length ) __snake_case : List[str] = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __snake_case : Optional[int] = max_length - minimum_tokens __snake_case : List[Any] = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: __snake_case : str = model_inputs["attention_mask"][:, -trim:] __snake_case : Any = model_inputs.pop("conversation" ) __snake_case : str = max_length __snake_case : Optional[int] = self.model.generate(**lowerCamelCase , **lowerCamelCase ) if self.model.config.is_encoder_decoder: __snake_case : List[Any] = 1 else: __snake_case : Union[str, Any] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __snake_case ( self : Any , lowerCamelCase : List[Any] , lowerCamelCase : Tuple=True ) -> Any: __snake_case : Optional[int] = model_outputs["output_ids"] __snake_case : Optional[Any] = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , ) __snake_case : Optional[int] = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowerCamelCase ) return conversation def __snake_case ( self : Optional[Any] , lowerCamelCase : Conversation ) -> Dict: __snake_case : Optional[Any] = self.tokenizer.eos_token_id __snake_case : Any = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) ) if len(lowerCamelCase ) > self.tokenizer.model_max_length: __snake_case : Tuple = input_ids[-self.tokenizer.model_max_length :] return input_ids
123
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """studio-ousia/luke-base""": """https://huggingface.co/studio-ousia/luke-base/resolve/main/config.json""", """studio-ousia/luke-large""": """https://huggingface.co/studio-ousia/luke-large/resolve/main/config.json""", } class A__ ( __lowercase ): lowercase = "luke" def __init__( self : Union[str, Any] , a : str=50_267 , a : Optional[int]=500_000 , a : Optional[int]=768 , a : Optional[Any]=256 , a : Optional[int]=12 , a : str=12 , a : int=3_072 , a : Tuple="gelu" , a : Any=0.1 , a : List[str]=0.1 , a : Union[str, Any]=512 , a : Union[str, Any]=2 , a : Tuple=0.0_2 , a : Any=1E-12 , a : List[Any]=True , a : Optional[int]=None , a : Tuple=1 , a : Optional[int]=0 , a : Optional[int]=2 , **a : int , ): super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : List[Any] = entity_vocab_size lowerCAmelCase__ : Optional[int] = hidden_size lowerCAmelCase__ : Optional[Any] = entity_emb_size lowerCAmelCase__ : List[Any] = num_hidden_layers lowerCAmelCase__ : Optional[int] = num_attention_heads lowerCAmelCase__ : Any = hidden_act lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : Dict = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : Tuple = max_position_embeddings lowerCAmelCase__ : Optional[Any] = type_vocab_size lowerCAmelCase__ : int = initializer_range lowerCAmelCase__ : List[str] = layer_norm_eps lowerCAmelCase__ : Dict = use_entity_aware_attention lowerCAmelCase__ : Optional[int] = classifier_dropout
365
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) lowerCamelCase__ = { """configuration_speech_to_text""": ["""SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Speech2TextConfig"""], """processing_speech_to_text""": ["""Speech2TextProcessor"""], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""Speech2TextTokenizer"""] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""Speech2TextFeatureExtractor"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSpeech2TextForConditionalGeneration""", """TFSpeech2TextModel""", """TFSpeech2TextPreTrainedModel""", ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Speech2TextForConditionalGeneration""", """Speech2TextModel""", """Speech2TextPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
307
0
'''simple docstring''' import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( lowerCAmelCase_ , unittest.TestCase ): A_ = LayoutLMTokenizer A_ = LayoutLMTokenizerFast A_ = True A_ = True def __UpperCAmelCase ( self ): '''simple docstring''' super().setUp() __a : List[Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __a : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __UpperCAmelCase ( self , **__a ): '''simple docstring''' return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **__a ) def __UpperCAmelCase ( self , __a ): '''simple docstring''' __a : Tuple = 'UNwant\u00E9d,running' __a : Tuple = 'unwanted, running' return input_text, output_text def __UpperCAmelCase ( self ): '''simple docstring''' __a : Any = self.tokenizer_class(self.vocab_file ) __a : Any = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(__a , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__a ) , [7, 4, 5, 10, 8, 9] ) def __UpperCAmelCase ( self ): '''simple docstring''' pass
27
'''simple docstring''' import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): @property def __UpperCAmelCase ( self ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = ort.SessionOptions() __a : Dict = False return options def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) __a : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) __a : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy' ) # using the PNDM scheduler by default __a : str = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='onnx' , safety_checker=__a , feature_extractor=__a , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__a ) __a : Tuple = 'A red cat sitting on a park bench' __a : int = np.random.RandomState(0 ) __a : Tuple = pipe( prompt=__a , image=__a , mask_image=__a , strength=0.75 , guidance_scale=7.5 , num_inference_steps=15 , generator=__a , output_type='np' , ) __a : Tuple = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
27
1
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def __a ( __lowerCamelCase, __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase, keep_in_memory=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = tmp_path / "cache" UpperCAmelCase_ : Optional[int] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : int = features.copy() if features else default_expected_features UpperCAmelCase_ : Optional[Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Any = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_3": "float64", "col_1": "string", "col_2": "int64"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = tmp_path / "cache" UpperCAmelCase_ : Dict = {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase_ : Any = features.copy() if features else default_expected_features UpperCAmelCase_ : Dict = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def __a ( __lowerCamelCase, __lowerCamelCase ): # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase_ : str = {"col_2": "int64", "col_3": "float64", "col_1": "string"} UpperCAmelCase_ : str = features.copy() UpperCAmelCase_ : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Any = tmp_path / "cache" UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Dict = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase, split=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type", [str, list] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if issubclass(__lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Dict = jsonl_path elif issubclass(__lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[Any] = [jsonl_path] UpperCAmelCase_ : str = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=("train",) ): assert isinstance(__lowerCamelCase, __lowerCamelCase ) for split in splits: UpperCAmelCase_ : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = tmp_path / "cache" UpperCAmelCase_ : Optional[Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase_ : List[str] = JsonDatasetReader({"train": jsonl_path}, cache_dir=__lowerCamelCase, keep_in_memory=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Any = features.copy() if features else default_expected_features UpperCAmelCase_ : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Optional[Any] = JsonDatasetReader({"train": jsonl_path}, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if split: UpperCAmelCase_ : Dict = {split: jsonl_path} else: UpperCAmelCase_ : Any = "train" UpperCAmelCase_ : Union[str, Any] = {"train": jsonl_path, "test": jsonl_path} UpperCAmelCase_ : Any = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase, splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def __a ( __lowerCamelCase ): return json.load(__lowerCamelCase ) def __a ( __lowerCamelCase ): return [json.loads(__lowerCamelCase ) for line in buffer] class A_ : '''simple docstring''' @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ ).write() buffer.seek(0 ) UpperCAmelCase_ : Dict = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ ).write() buffer.seek(0 ) UpperCAmelCase_ : Optional[Any] = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase_ : Optional[Any] = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase_ : List[Any] = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" with pytest.raises(lowercase_ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , num_proc=0 ) @pytest.mark.parametrize("compression, extension" , [("gzip", "gz"), ("bz2", "bz2"), ("xz", "xz")] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = tmp_path_factory.mktemp("data" ) / F"""test.json.{extension}""" UpperCAmelCase_ : str = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowercase_ , lowercase_ , compression=lowercase_ ).write() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: UpperCAmelCase_ : List[str] = f.read() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: UpperCAmelCase_ : Dict = f.read() assert exported_content == original_content
23
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
1
A : List[str] = { "km/h": 1.0, "m/s": 3.6, "mph": 1.6_0_9_3_4_4, "knot": 1.8_5_2, } A : Optional[int] = { "km/h": 1.0, "m/s": 0.2_7_7_7_7_7_7_7_8, "mph": 0.6_2_1_3_7_1_1_9_2, "knot": 0.5_3_9_9_5_6_8_0_3, } def lowercase_ ( _A : float , _A : str , _A : str ): """simple docstring""" if unit_to not in speed_chart or unit_from not in speed_chart_inverse: lowerCamelCase__ : List[Any] = ( F"Incorrect 'from_type' or 'to_type' value: {unit_from!r}, {unit_to!r}\n" F"Valid values are: {', '.join(__lowerCAmelCase )}" ) raise ValueError(__lowerCAmelCase ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
184
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def lowerCamelCase__ ( __lowerCAmelCase : str = "" ): """simple docstring""" lowerCAmelCase_ = url or "https://www.imdb.com/chart/top/?ref_=nv_mv_250" lowerCAmelCase_ = BeautifulSoup(requests.get(__lowerCAmelCase ).text , "html.parser" ) lowerCAmelCase_ = soup.find_all("td" , attrs="titleColumn" ) lowerCAmelCase_ = soup.find_all("td" , class_="ratingColumn imdbRating" ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(__lowerCAmelCase , __lowerCAmelCase ) } def lowerCamelCase__ ( __lowerCAmelCase : str = "IMDb_Top_250_Movies.csv" ): """simple docstring""" lowerCAmelCase_ = get_imdb_top_aaa_movies() with open(__lowerCAmelCase , "w" , newline="" ) as out_file: lowerCAmelCase_ = csv.writer(__lowerCAmelCase ) writer.writerow(["Movie title", "IMDb rating"] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
231
0
from string import ascii_uppercase _SCREAMING_SNAKE_CASE = {str(ord(c) - 55): c for c in ascii_uppercase} def snake_case ( snake_case__ :int , snake_case__ :int) -> List[str]: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): raise TypeError("""int() can't convert non-string with explicit base""") if num < 0: raise ValueError("""parameter must be positive int""") if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): raise TypeError("""'str' object cannot be interpreted as an integer""") if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): raise TypeError("""'float' object cannot be interpreted as an integer""") if base in (0, 1): raise ValueError("""base must be >= 2""") if base > 36: raise ValueError("""base must be <= 36""") _A = """""" _A = 0 _A = 0 while div != 1: _A , _A = divmod(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) if base >= 11 and 9 < mod < 36: _A = ALPHABET_VALUES[str(SCREAMING_SNAKE_CASE__)] else: _A = str(SCREAMING_SNAKE_CASE__) new_value += actual_value _A = num // base _A = div if div == 0: return str(new_value[::-1]) elif div == 1: new_value += str(SCREAMING_SNAKE_CASE__) return str(new_value[::-1]) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1_000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
367
import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = [ 'word_embeddings_layernorm.weight', 'word_embeddings_layernorm.bias', 'input_layernorm.weight', 'input_layernorm.bias', 'post_attention_layernorm.weight', 'post_attention_layernorm.bias', 'self_attention.dense.bias', 'mlp.dense_4h_to_h.bias', 'ln_f.weight', 'ln_f.bias', ] _SCREAMING_SNAKE_CASE = [ 'mlp.dense_4h_to_h.weight', 'self_attention.dense.weight', ] def snake_case ( snake_case__ :List[str] , snake_case__ :Dict) -> str: _A = { """word_embeddings.weight""": """word_embeddings.weight""", """word_embeddings.norm.weight""": """word_embeddings_layernorm.weight""", """word_embeddings.norm.bias""": """word_embeddings_layernorm.bias""", """weight""": """ln_f.weight""", """bias""": """ln_f.bias""", } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks _A = int(re.match(R""".*layer_(\d*).*""" , snake_case__)[1]) layer_number -= 3 return F'''h.{layer_number}.''' + key def snake_case ( snake_case__ :Tuple) -> int: if dtype == torch.bool: return 1 / 8 _A = re.search(R"""[^\d](\d+)$""" , str(snake_case__)) if bit_search is None: raise ValueError(F'''`dtype` is not a valid dtype: {dtype}.''') _A = int(bit_search.groups()[0]) return bit_size // 8 def snake_case ( snake_case__ :Dict , snake_case__ :Any , snake_case__ :Union[str, Any] , snake_case__ :Tuple , snake_case__ :List[Any]) -> List[str]: # Construct model if bloom_config_file == "": _A = BloomConfig() else: _A = BloomConfig.from_json_file(snake_case__) if shard_model: _A = os.listdir(snake_case__) _A = sorted(filter(lambda snake_case__: s.startswith("""layer""") and "model_00" in s , snake_case__)) _A = {"""weight_map""": {}, """metadata""": {}} _A = 0 _A = None _A = BloomConfig() for j, file in enumerate(snake_case__): print("""Processing file: {}""".format(snake_case__)) _A = None for i in range(snake_case__): # load all TP files _A = file.replace("""model_00""" , F'''model_0{i}''') _A = torch.load(os.path.join(snake_case__ , snake_case__) , map_location="""cpu""") # Rename keys in the transformers names _A = list(temp.keys()) for key in keys: _A = temp.pop(snake_case__) if tensors is None: _A = temp else: for key in tensors.keys(): if any(key.endswith(snake_case__) for end in WEIGHTS_TO_AVERAGE_ENDSWITH): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel _A = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN) else 0 # We concatenate these weights accross TP ranks _A = torch.cat([tensors[key], temp[key]] , dim=snake_case__) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(snake_case__) for end in WEIGHTS_TO_AVERAGE_ENDSWITH): _A = tensors[key] / pretraining_tp torch.save( snake_case__ , os.path.join( snake_case__ , """pytorch_model_{}-of-{}.bin""".format(str(j + 1).zfill(5) , str(len(snake_case__)).zfill(5)) , ) , ) for key in tensors.keys(): _A = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype) if key not in index_dict["weight_map"]: _A = """pytorch_model_{}-of-{}.bin""".format( str(j + 1).zfill(5) , str(len(snake_case__)).zfill(5)) _A = BloomConfig() _A = pytorch_dump_folder_path + """/""" + CONFIG_NAME _A = total_size with open(snake_case__ , """w""" , encoding="""utf-8""") as f: f.write(config.to_json_string()) with open(os.path.join(snake_case__ , WEIGHTS_NAME + """.index.json""") , """w""" , encoding="""utf-8""") as f: _A = json.dumps(snake_case__ , indent=2 , sort_keys=snake_case__) + """\n""" f.write(snake_case__) else: _A = BloomModel(snake_case__) _A = os.listdir(snake_case__) _A = sorted(filter(lambda snake_case__: s.startswith("""layer""") and "model_00" in s , snake_case__)) _A = None for i, file in enumerate(snake_case__): _A = None for i in range(snake_case__): # load all TP files _A = file.replace("""model_00""" , F'''model_0{i}''') _A = torch.load(os.path.join(snake_case__ , snake_case__) , map_location="""cpu""") # Rename keys in the transformers names _A = list(temp.keys()) for key in keys: _A = temp.pop(snake_case__) if tensors is None: _A = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(snake_case__) for end in WEIGHTS_TO_AVERAGE_ENDSWITH): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel _A = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN) else 0 # We concatenate these weights accross TP ranks _A = torch.cat([tensors[key], temp[key]] , dim=snake_case__) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(snake_case__) for end in WEIGHTS_TO_AVERAGE_ENDSWITH): _A = tensors[key] / pretraining_tp _A = model.load_state_dict(snake_case__ , strict=snake_case__) assert not other_keys.unexpected_keys, F'''The keys {other_keys.unexpected_keys} are unexpected''' if missing_keys is None: _A = set(other_keys.missing_keys) else: _A = missing_keys.intersection(set(other_keys.missing_keys)) assert not missing_keys, F'''The keys {missing_keys} are missing''' # Save pytorch-model os.makedirs(snake_case__ , exist_ok=snake_case__) _A = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME _A = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}''') if config.torch_dtype is not None: _A = model.to(config.torch_dtype) torch.save(model.state_dict() , snake_case__) print(F'''Save configuration file to {pytorch_config_dump_path}''') with open(snake_case__ , """w""" , encoding="""utf-8""") as f: f.write(config.to_json_string()) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '--bloom_checkpoint_path', default=None, type=str, required=True, help='Path to the Megatron-LM checkpoint path.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--bloom_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--shard_model', action='store_true', help='An optional setting to shard the output model \nThis enables sharding the converted checkpoint', ) parser.add_argument( '--pretraining_tp', default=4, type=int, help='Pretraining TP rank that has been used when training the model in Megatron-LM \n', ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
81
0
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class _lowercase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' _SCREAMING_SNAKE_CASE : Any = None _SCREAMING_SNAKE_CASE : Any = None @property def a ( self : List[Any] ) -> str: return self.feat_extract_tester.prepare_feat_extract_dict() def a ( self : int ) -> int: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_A , """feature_size""" ) ) self.assertTrue(hasattr(_A , """sampling_rate""" ) ) self.assertTrue(hasattr(_A , """padding_value""" ) ) def a ( self : Tuple ) -> Tuple: __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common() __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_A ) == len(_A ) for x, y in zip(_A , processed_features[input_name] ) ) ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_A ) __lowerCAmelCase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowerCAmelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowerCAmelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def a ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_A ) __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowerCAmelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowerCAmelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def a ( self : Tuple ) -> Union[str, Any]: __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_A ) __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) __lowerCAmelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowerCAmelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def a ( self : Tuple , SCREAMING_SNAKE_CASE__ : str=False ) -> Tuple: def _inputs_have_equal_length(SCREAMING_SNAKE_CASE__ : List[str] ): __lowerCAmelCase = len(input[0] ) for input_slice in input[1:]: if len(_A ) != length: return False return True def _inputs_are_equal(SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ): if len(_A ) != len(_A ): return False for input_slice_a, input_slice_a in zip(_A , _A ): if not np.allclose(np.asarray(_A ) , np.asarray(_A ) , atol=1e-3 ): return False return True __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_A ) __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) __lowerCAmelCase = self.feat_extract_tester.seq_length_diff __lowerCAmelCase = self.feat_extract_tester.max_seq_length + pad_diff __lowerCAmelCase = self.feat_extract_tester.min_seq_length __lowerCAmelCase = self.feat_extract_tester.batch_size __lowerCAmelCase = self.feat_extract_tester.feature_size # test padding for List[int] + numpy __lowerCAmelCase = feat_extract.pad(_A , padding=_A ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""np""" ) __lowerCAmelCase = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_A ): feat_extract.pad(_A , padding="""max_length""" )[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=_A , return_tensors="""np""" ) __lowerCAmelCase = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_A ) ) self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertTrue(_inputs_are_equal(_A , _A ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy __lowerCAmelCase = feat_extract.pad(_A , pad_to_multiple_of=10 ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , pad_to_multiple_of=10 ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_A ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , pad_to_multiple_of=10 , max_length=_A , return_tensors="""np""" , ) __lowerCAmelCase = input_a[input_name] self.assertTrue(all(len(_A ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_A , _A ) ) __lowerCAmelCase = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_A ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct __lowerCAmelCase = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1e-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1e-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1e-3 ) def a ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[int]=False ) -> Tuple: def _inputs_have_equal_length(SCREAMING_SNAKE_CASE__ : List[Any] ): __lowerCAmelCase = len(input[0] ) for input_slice in input[1:]: if len(_A ) != length: return False return True def _inputs_are_equal(SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] ): if len(_A ) != len(_A ): return False for input_slice_a, input_slice_a in zip(_A , _A ): if not np.allclose(np.asarray(_A ) , np.asarray(_A ) , atol=1e-3 ): return False return True __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common(numpify=_A ) __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) # truncate to smallest __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=_A ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) __lowerCAmelCase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertFalse(_inputs_have_equal_length(_A ) ) # truncate to smallest with np __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=_A , ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) __lowerCAmelCase = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_A ) ) # truncate to middle __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_A , return_tensors="""np""" , ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=_A ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) __lowerCAmelCase = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertTrue(_inputs_are_equal(_A , _A ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_A ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_A ): feat_extract.pad(_A , truncation=_A )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_A ): feat_extract.pad(_A , padding="""longest""" , truncation=_A )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_A ): feat_extract.pad(_A , padding="""longest""" , truncation=_A )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_A ): feat_extract.pad(_A , padding="""max_length""" , truncation=_A )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy __lowerCAmelCase = 12 __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_A , truncation=_A , ) __lowerCAmelCase = input_a[input_name] __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=_A , ) __lowerCAmelCase = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of __lowerCAmelCase = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: __lowerCAmelCase = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_A ) ) self.assertFalse(_inputs_have_equal_length(_A ) ) def a ( self : Any ) -> List[str]: self._check_padding(numpify=_A ) def a ( self : Optional[Any] ) -> int: self._check_padding(numpify=_A ) def a ( self : List[Any] ) -> Dict: self._check_truncation(numpify=_A ) def a ( self : Tuple ) -> Any: self._check_truncation(numpify=_A ) @require_torch def a ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common() __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""np""" )[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1e-2 ) @require_tf def a ( self : List[Any] ) -> Tuple: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common() __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""np""" )[input_name] __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1e-2 ) def a ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = self.feat_extract_dict __lowerCAmelCase = True __lowerCAmelCase = self.feature_extraction_class(**_A ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common() __lowerCAmelCase = [len(_A ) for x in speech_inputs] __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) __lowerCAmelCase = feat_extract.pad(_A , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _A ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , _A ) def a ( self : int ) -> Dict: __lowerCAmelCase = self.feat_extract_dict __lowerCAmelCase = True __lowerCAmelCase = self.feature_extraction_class(**_A ) __lowerCAmelCase = self.feat_extract_tester.prepare_inputs_for_common() __lowerCAmelCase = [len(_A ) for x in speech_inputs] __lowerCAmelCase = feat_extract.model_input_names[0] __lowerCAmelCase = BatchFeature({input_name: speech_inputs} ) __lowerCAmelCase = min(_A ) __lowerCAmelCase = feat_extract.pad( _A , padding="""max_length""" , max_length=_A , truncation=_A , return_tensors="""np""" ) self.assertIn("""attention_mask""" , _A ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
229
# limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( "pipelines_utils", "0.22.0", "Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.", standard_warn=False, stacklevel=3, )
299
0
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ): """simple docstring""" if radian_mode: return [magnitude * cos(_SCREAMING_SNAKE_CASE ), magnitude * sin(_SCREAMING_SNAKE_CASE )] return [magnitude * cos(radians(_SCREAMING_SNAKE_CASE ) ), magnitude * sin(radians(_SCREAMING_SNAKE_CASE ) )] def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10**-1 ): """simple docstring""" UpperCamelCase = cross(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = sum(_SCREAMING_SNAKE_CASE ) return abs(_SCREAMING_SNAKE_CASE ) < eps if __name__ == "__main__": # Test to check if it works lowerCAmelCase__ = array( [ polar_force(7_18.4, 180 - 30), polar_force(8_79.54, 45), polar_force(100, -90), ] ) lowerCAmelCase__ = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg lowerCAmelCase__ = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) lowerCAmelCase__ = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg lowerCAmelCase__ = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) lowerCAmelCase__ = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
244
"""simple docstring""" from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING lowerCAmelCase__ = logging.get_logger(__name__) @add_end_docstrings(_lowercase ) class _lowerCamelCase ( _lowercase ): def __init__(self , **__a ) -> Optional[int]: super().__init__(**__a ) requires_backends(self , "vision" ) requires_backends(self , "torch" ) if self.framework != "pt": raise ValueError(F"The {self.__class__} is only available in PyTorch." ) self.check_model_type(__a ) def snake_case_ (self , **__a ) -> List[Any]: UpperCamelCase = {} UpperCamelCase = {} UpperCamelCase = {} # preprocess args if "points_per_batch" in kwargs: UpperCamelCase = kwargs["points_per_batch"] if "points_per_crop" in kwargs: UpperCamelCase = kwargs["points_per_crop"] if "crops_n_layers" in kwargs: UpperCamelCase = kwargs["crops_n_layers"] if "crop_overlap_ratio" in kwargs: UpperCamelCase = kwargs["crop_overlap_ratio"] if "crop_n_points_downscale_factor" in kwargs: UpperCamelCase = kwargs["crop_n_points_downscale_factor"] # postprocess args if "pred_iou_thresh" in kwargs: UpperCamelCase = kwargs["pred_iou_thresh"] if "stability_score_offset" in kwargs: UpperCamelCase = kwargs["stability_score_offset"] if "mask_threshold" in kwargs: UpperCamelCase = kwargs["mask_threshold"] if "stability_score_thresh" in kwargs: UpperCamelCase = kwargs["stability_score_thresh"] if "crops_nms_thresh" in kwargs: UpperCamelCase = kwargs["crops_nms_thresh"] if "output_rle_mask" in kwargs: UpperCamelCase = kwargs["output_rle_mask"] if "output_bboxes_mask" in kwargs: UpperCamelCase = kwargs["output_bboxes_mask"] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__(self , __a , *__a , __a=None , __a=None , **__a ) -> str: return super().__call__(__a , *__a , num_workers=__a , batch_size=__a , **__a ) def snake_case_ (self , __a , __a=64 , __a = 0 , __a = 5_12 / 15_00 , __a = 32 , __a = 1 , ) -> List[str]: UpperCamelCase = load_image(__a ) UpperCamelCase = self.image_processor.size["longest_edge"] UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self.image_processor.generate_crop_boxes( __a , __a , __a , __a , __a , __a ) UpperCamelCase = self.image_processor(images=__a , return_tensors="pt" ) with self.device_placement(): if self.framework == "pt": UpperCamelCase = self.get_inference_context() with inference_context(): UpperCamelCase = self._ensure_tensor_on_device(__a , device=self.device ) UpperCamelCase = self.model.get_image_embeddings(model_inputs.pop("pixel_values" ) ) UpperCamelCase = image_embeddings UpperCamelCase = grid_points.shape[1] UpperCamelCase = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( "Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. " "To return all points at once, set points_per_batch to None" ) for i in range(0 , __a , __a ): UpperCamelCase = grid_points[:, i : i + points_per_batch, :, :] UpperCamelCase = input_labels[:, i : i + points_per_batch] UpperCamelCase = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def snake_case_ (self , __a , __a=0.88 , __a=0.95 , __a=0 , __a=1 , ) -> int: UpperCamelCase = model_inputs.pop("input_boxes" ) UpperCamelCase = model_inputs.pop("is_last" ) UpperCamelCase = model_inputs.pop("original_sizes" ).tolist() UpperCamelCase = model_inputs.pop("reshaped_input_sizes" ).tolist() UpperCamelCase = self.model(**__a ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks UpperCamelCase = model_outputs["pred_masks"] UpperCamelCase = self.image_processor.post_process_masks( __a , __a , __a , __a , binarize=__a ) UpperCamelCase = model_outputs["iou_scores"] UpperCamelCase , UpperCamelCase , UpperCamelCase = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __a , __a , __a , __a , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def snake_case_ (self , __a , __a=False , __a=False , __a=0.7 , ) -> Optional[int]: UpperCamelCase = [] UpperCamelCase = [] UpperCamelCase = [] for model_output in model_outputs: all_scores.append(model_output.pop("iou_scores" ) ) all_masks.extend(model_output.pop("masks" ) ) all_boxes.append(model_output.pop("boxes" ) ) UpperCamelCase = torch.cat(__a ) UpperCamelCase = torch.cat(__a ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self.image_processor.post_process_for_mask_generation( __a , __a , __a , __a ) UpperCamelCase = defaultdict(__a ) for output in model_outputs: for k, v in output.items(): extra[k].append(__a ) UpperCamelCase = {} if output_rle_mask: UpperCamelCase = rle_mask if output_bboxes_mask: UpperCamelCase = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
244
1
"""simple docstring""" from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def __magic_name__ ( __snake_case : Any , __snake_case : Any , __snake_case : Dict=1E-12 ) -> Dict: lowercase : List[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__snake_case , axis=1 ) , a_min=__snake_case ) ).T lowercase : str = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__snake_case , axis=1 ) , a_min=__snake_case ) ).T return jnp.matmul(__snake_case , norm_emb_a.T ) class a__ ( nn.Module ): __lowerCAmelCase = 42 __lowerCAmelCase = jnp.floataa def __magic_name__ ( self ): lowercase : int = FlaxCLIPVisionModule(self.config.vision_config ) lowercase : Union[str, Any] = nn.Dense(self.config.projection_dim , use_bias=_a , dtype=self.dtype ) lowercase : Optional[Any] = self.param("concept_embeds" , jax.nn.initializers.ones , (17, self.config.projection_dim) ) lowercase : int = self.param( "special_care_embeds" , jax.nn.initializers.ones , (3, self.config.projection_dim) ) lowercase : str = self.param("concept_embeds_weights" , jax.nn.initializers.ones , (17,) ) lowercase : Dict = self.param("special_care_embeds_weights" , jax.nn.initializers.ones , (3,) ) def __call__( self , _a ): lowercase : Any = self.vision_model(_a )[1] lowercase : Union[str, Any] = self.visual_projection(_a ) lowercase : List[Any] = jax_cosine_distance(_a , self.special_care_embeds ) lowercase : str = jax_cosine_distance(_a , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs lowercase : List[str] = 0.0 lowercase : Union[str, Any] = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment lowercase : Tuple = jnp.round(_a , 3 ) lowercase : List[str] = jnp.any(special_scores > 0 , axis=1 , keepdims=_a ) # Use a lower threshold if an image has any special care concept lowercase : Tuple = is_special_care * 0.0_1 lowercase : Optional[Any] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment lowercase : Union[str, Any] = jnp.round(_a , 3 ) lowercase : Tuple = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class a__ ( a_ ): __lowerCAmelCase = CLIPConfig __lowerCAmelCase = """clip_input""" __lowerCAmelCase = FlaxStableDiffusionSafetyCheckerModule def __init__( self , _a , _a = None , _a = 0 , _a = jnp.floataa , _a = True , **_a , ): if input_shape is None: lowercase : List[Any] = (1, 224, 224, 3) lowercase : Union[str, Any] = self.module_class(config=_a , dtype=_a , **_a ) super().__init__(_a , _a , input_shape=_a , seed=_a , dtype=_a , _do_init=_do_init ) def __magic_name__ ( self , _a , _a , _a = None ): # init input tensor lowercase : Tuple = jax.random.normal(_a , _a ) lowercase , lowercase : int = jax.random.split(_a ) lowercase : List[str] = {"params": params_rng, "dropout": dropout_rng} lowercase : Union[str, Any] = self.module.init(_a , _a )["params"] return random_params def __call__( self , _a , _a = None , ): lowercase : str = jnp.transpose(_a , (0, 2, 3, 1) ) return self.module.apply( {"params": params or self.params} , jnp.array(_a , dtype=jnp.floataa ) , rngs={} , )
202
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin _A : int = """ Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] """ class a__ ( unittest.TestCase, a_ ): def __magic_name__ ( self ): lowercase : Tuple = load_tool("text-question-answering" ) self.tool.setup() lowercase : Dict = load_tool("text-question-answering" , remote=_a ) def __magic_name__ ( self ): lowercase : str = self.tool(_a , "What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : Union[str, Any] = self.remote_tool(_a , "What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : int = self.tool(text=_a , question="What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : Optional[Any] = self.remote_tool(text=_a , question="What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" )
202
1
import inspect import unittest class __A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ (self ): try: import diffusers # noqa: F401 except ImportError: assert False def UpperCAmelCase_ (self ): import diffusers from diffusers.dependency_versions_table import deps UpperCamelCase__ = inspect.getmembers(SCREAMING_SNAKE_CASE_ , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": UpperCamelCase__ = """k-diffusion""" elif backend == "invisible_watermark": UpperCamelCase__ = """invisible-watermark""" assert backend in deps, F"{backend} is not in the deps table!"
178
import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class __A( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ (self ): UpperCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) UpperCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = -1 UpperCamelCase__ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=10 , do_sample=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: UpperCamelCase__ = TextStreamer(SCREAMING_SNAKE_CASE_ ) model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=10 , do_sample=SCREAMING_SNAKE_CASE_ , streamer=SCREAMING_SNAKE_CASE_ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer UpperCamelCase__ = cs.out[:-1] self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) UpperCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = -1 UpperCamelCase__ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=10 , do_sample=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.decode(greedy_ids[0] ) UpperCamelCase__ = TextIteratorStreamer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} UpperCamelCase__ = Thread(target=model.generate , kwargs=SCREAMING_SNAKE_CASE_ ) thread.start() UpperCamelCase__ = """""" for new_text in streamer: streamer_text += new_text self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): UpperCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) UpperCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = -1 UpperCamelCase__ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=10 , do_sample=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = greedy_ids[:, input_ids.shape[1] :] UpperCamelCase__ = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: UpperCamelCase__ = TextStreamer(SCREAMING_SNAKE_CASE_ , skip_prompt=SCREAMING_SNAKE_CASE_ ) model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=10 , do_sample=SCREAMING_SNAKE_CASE_ , streamer=SCREAMING_SNAKE_CASE_ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer UpperCamelCase__ = cs.out[:-1] self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self ): # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them UpperCamelCase__ = AutoTokenizer.from_pretrained("""distilgpt2""" ) UpperCamelCase__ = AutoModelForCausalLM.from_pretrained("""distilgpt2""" ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = -1 UpperCamelCase__ = torch.ones((1, 5) , device=SCREAMING_SNAKE_CASE_ ).long() * model.config.bos_token_id with CaptureStdout() as cs: UpperCamelCase__ = TextStreamer(SCREAMING_SNAKE_CASE_ , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) model.generate(SCREAMING_SNAKE_CASE_ , max_new_tokens=1 , do_sample=SCREAMING_SNAKE_CASE_ , streamer=SCREAMING_SNAKE_CASE_ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token UpperCamelCase__ = cs.out[:-1] # Remove the final "\n" UpperCamelCase__ = tokenizer(SCREAMING_SNAKE_CASE_ , return_tensors="""pt""" ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def UpperCAmelCase_ (self ): UpperCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) UpperCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = -1 UpperCamelCase__ = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = TextIteratorStreamer(SCREAMING_SNAKE_CASE_ , timeout=0.001 ) UpperCamelCase__ = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} UpperCamelCase__ = Thread(target=model.generate , kwargs=SCREAMING_SNAKE_CASE_ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = """""" for new_text in streamer: streamer_text += new_text
178
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss lowerCAmelCase__ : Dict = pytest.mark.integration @require_faiss class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCamelCase__ ) for x in np.arange(30 ).tolist()]} ) return dset def __lowerCAmelCase ( self : List[Any] ): import faiss UpperCAmelCase__ = self._create_dummy_dataset() UpperCAmelCase__ = dset.map( lambda lowerCamelCase__ ,lowerCamelCase__ : {"vecs": i * np.ones(5 ,dtype=np.floataa )} ,with_indices=lowerCamelCase__ ,keep_in_memory=lowerCamelCase__ ) UpperCAmelCase__ = dset.add_faiss_index('vecs' ,batch_size=100 ,metric_type=faiss.METRIC_INNER_PRODUCT ) UpperCAmelCase__ , UpperCAmelCase__ = dset.get_nearest_examples('vecs' ,np.ones(5 ,dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] ,'my_name-train_29' ) dset.drop_index('vecs' ) def __lowerCAmelCase ( self : List[str] ): import faiss UpperCAmelCase__ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 ,1 ) ,index_name='vecs' ,batch_size=100 ,metric_type=faiss.METRIC_INNER_PRODUCT ,) UpperCAmelCase__ , UpperCAmelCase__ = dset.get_nearest_examples('vecs' ,np.ones(5 ,dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] ,'my_name-train_29' ) def __lowerCAmelCase ( self : List[Any] ): import faiss UpperCAmelCase__ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 ,1 ) ,index_name='vecs' ,metric_type=faiss.METRIC_INNER_PRODUCT ,) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCamelCase__ ) as tmp_file: dset.save_faiss_index('vecs' ,tmp_file.name ) dset.load_faiss_index('vecs2' ,tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase__ , UpperCAmelCase__ = dset.get_nearest_examples('vecs2' ,np.ones(5 ,dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] ,'my_name-train_29' ) def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 ,1 ) ,index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCamelCase__ ,partial(dset.get_nearest_examples ,'vecs2' ,np.ones(5 ,dtype=np.floataa ) ) ) def __lowerCAmelCase ( self : str ): from elasticsearch import Elasticsearch UpperCAmelCase__ = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: UpperCAmelCase__ = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) UpperCAmelCase__ = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} UpperCAmelCase__ = Elasticsearch() dset.add_elasticsearch_index('filename' ,es_client=lowerCamelCase__ ) UpperCAmelCase__ , UpperCAmelCase__ = dset.get_nearest_examples('filename' ,'my_name-train_29' ) self.assertEqual(examples['filename'][0] ,'my_name-train_29' ) @require_faiss class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): import faiss UpperCAmelCase__ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 ,dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal ,5 ) index.add_vectors(np.zeros((5, 5) ,dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal ,10 ) # single query UpperCAmelCase__ = np.zeros(5 ,dtype=np.floataa ) UpperCAmelCase__ = 1 UpperCAmelCase__ , UpperCAmelCase__ = index.search(lowerCamelCase__ ) self.assertRaises(lowerCamelCase__ ,index.search ,query.reshape(-1 ,1 ) ) self.assertGreater(scores[0] ,0 ) self.assertEqual(indices[0] ,1 ) # batched queries UpperCAmelCase__ = np.eye(5 ,dtype=np.floataa )[::-1] UpperCAmelCase__ , UpperCAmelCase__ = index.search_batch(lowerCamelCase__ ) self.assertRaises(lowerCamelCase__ ,index.search_batch ,queries[0] ) UpperCAmelCase__ = [scores[0] for scores in total_scores] UpperCAmelCase__ = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCamelCase__ ) ,0 ) self.assertListEqual([4, 3, 2, 1, 0] ,lowerCamelCase__ ) def __lowerCAmelCase ( self : Dict ): import faiss UpperCAmelCase__ = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 ,dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index ,faiss.IndexFlat ) UpperCAmelCase__ = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 ,dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index ,faiss.IndexLSH ) with self.assertRaises(lowerCamelCase__ ): UpperCAmelCase__ = FaissIndex(string_factory='Flat' ,custom_index=faiss.IndexFlat(5 ) ) def __lowerCAmelCase ( self : str ): import faiss UpperCAmelCase__ = faiss.IndexFlat(5 ) UpperCAmelCase__ = FaissIndex(custom_index=lowerCamelCase__ ) index.add_vectors(np.eye(5 ,dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index ,faiss.IndexFlat ) def __lowerCAmelCase ( self : List[Any] ): import faiss UpperCAmelCase__ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 ,dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCamelCase__ ) as tmp_file: index.save(tmp_file.name ) UpperCAmelCase__ = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase__ = np.zeros(5 ,dtype=np.floataa ) UpperCAmelCase__ = 1 UpperCAmelCase__ , UpperCAmelCase__ = index.search(lowerCamelCase__ ) self.assertGreater(scores[0] ,0 ) self.assertEqual(indices[0] ,1 ) @require_faiss def a_ ( lowerCamelCase ): import faiss UpperCAmelCase__ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) UpperCAmelCase__ = 'index.faiss' UpperCAmelCase__ = f'''mock://{index_name}''' index.save(lowerCamelCase , storage_options=mockfs.storage_options ) UpperCAmelCase__ = FaissIndex.load(lowerCamelCase , storage_options=mockfs.storage_options ) UpperCAmelCase__ = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase__ = 1 UpperCAmelCase__ , UpperCAmelCase__ = index.search(lowerCamelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : Any ): from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: UpperCAmelCase__ = Elasticsearch() UpperCAmelCase__ = {'acknowledged': True} UpperCAmelCase__ = ElasticSearchIndex(es_client=lowerCamelCase__ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query UpperCAmelCase__ = 'foo' UpperCAmelCase__ = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} UpperCAmelCase__ , UpperCAmelCase__ = index.search(lowerCamelCase__ ) self.assertEqual(scores[0] ,1 ) self.assertEqual(indices[0] ,0 ) # single query with timeout UpperCAmelCase__ = 'foo' UpperCAmelCase__ = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} UpperCAmelCase__ , UpperCAmelCase__ = index.search(lowerCamelCase__ ,request_timeout=30 ) self.assertEqual(scores[0] ,1 ) self.assertEqual(indices[0] ,0 ) # batched queries UpperCAmelCase__ = ['foo', 'bar', 'foobar'] UpperCAmelCase__ = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} UpperCAmelCase__ , UpperCAmelCase__ = index.search_batch(lowerCamelCase__ ) UpperCAmelCase__ = [scores[0] for scores in total_scores] UpperCAmelCase__ = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCamelCase__ ) ,0 ) self.assertListEqual([1, 1, 1] ,lowerCamelCase__ ) # batched queries with timeout UpperCAmelCase__ = ['foo', 'bar', 'foobar'] UpperCAmelCase__ = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} UpperCAmelCase__ , UpperCAmelCase__ = index.search_batch(lowerCamelCase__ ,request_timeout=30 ) UpperCAmelCase__ = [scores[0] for scores in total_scores] UpperCAmelCase__ = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCamelCase__ ) ,0 ) self.assertListEqual([1, 1, 1] ,lowerCamelCase__ )
98
"""simple docstring""" lowerCamelCase_ = [ (1000, '''M'''), (900, '''CM'''), (500, '''D'''), (400, '''CD'''), (100, '''C'''), (90, '''XC'''), (50, '''L'''), (40, '''XL'''), (10, '''X'''), (9, '''IX'''), (5, '''V'''), (4, '''IV'''), (1, '''I'''), ] def snake_case ( A__ ): UpperCAmelCase_ : List[str] = {"I": 1, "V": 5, "X": 10, "L": 50, "C": 1_00, "D": 5_00, "M": 10_00} UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Tuple = 0 while place < len(A__ ): if (place + 1 < len(A__ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def snake_case ( A__ ): UpperCAmelCase_ : Union[str, Any] = [] for arabic, roman in ROMAN: ((UpperCAmelCase_) , (UpperCAmelCase_)) : str = divmod(A__ ,A__ ) result.append(roman * factor ) if number == 0: break return "".join(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
268
0
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->Tuple: """simple docstring""" if (ksize % 2) == 0: lowercase : List[str] = ksize + 1 lowercase : str = np.zeros((ksize, ksize), dtype=np.floataa ) # each value for y in range(A__ ): for x in range(A__ ): # distance from center lowercase : Optional[Any] = x - ksize // 2 lowercase : int = y - ksize // 2 # degree to radiant lowercase : Dict = theta / 180 * np.pi lowercase : str = np.cos(_theta ) lowercase : Tuple = np.sin(_theta ) # get kernel x lowercase : Tuple = cos_theta * px + sin_theta * py # get kernel y lowercase : int = -sin_theta * px + cos_theta * py # fill kernel lowercase : Optional[Any] = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image __a = imread('''../image_data/lena.jpg''') # turn image in gray scale value __a = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges __a = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 1_20, 1_50]: __a = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) __a = out / out.max() * 2_55 __a = out.astype(np.uinta) imshow('''Original''', gray) imshow('''Gabor filter with 20x20 mask and 6 directions''', out) waitKey(0)
358
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __SCREAMING_SNAKE_CASE ( A__ ): def __lowerCamelCase ( self ): lowercase : List[str] = tempfile.mkdtemp() lowercase : Any = 8 # DPR tok lowercase : List[str] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowercase : int = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(SCREAMING_SNAKE_CASE__ , exist_ok=SCREAMING_SNAKE_CASE__ ) lowercase : str = os.path.join(SCREAMING_SNAKE_CASE__ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowercase : int = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowercase : Optional[int] = dict(zip(SCREAMING_SNAKE_CASE__ , range(len(SCREAMING_SNAKE_CASE__ ) ) ) ) lowercase : str = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowercase : Union[str, Any] = {'''unk_token''': '''<unk>'''} lowercase : int = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(SCREAMING_SNAKE_CASE__ , exist_ok=SCREAMING_SNAKE_CASE__ ) lowercase : Optional[int] = os.path.join(SCREAMING_SNAKE_CASE__ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase : int = os.path.join(SCREAMING_SNAKE_CASE__ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(SCREAMING_SNAKE_CASE__ ) ) def __lowerCamelCase ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __lowerCamelCase ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __lowerCamelCase ( self ): shutil.rmtree(self.tmpdirname ) @require_tokenizers def __lowerCamelCase ( self ): lowercase : Optional[Any] = os.path.join(self.tmpdirname , '''rag_tokenizer''' ) lowercase : Dict = RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) lowercase : Dict = RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(SCREAMING_SNAKE_CASE__ ) rag_tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) lowercase : Dict = RagTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ , config=SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(new_rag_tokenizer.question_encoder , SCREAMING_SNAKE_CASE__ ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , SCREAMING_SNAKE_CASE__ ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __lowerCamelCase ( self ): lowercase : Dict = RagTokenizer.from_pretrained('''facebook/rag-token-nq''' ) lowercase : Optional[int] = [ '''who got the first nobel prize in physics''', '''when is the next deadpool movie being released''', '''which mode is used for short wave broadcast service''', '''who is the owner of reading football club''', '''when is the next scandal episode coming out''', '''when is the last time the philadelphia won the superbowl''', '''what is the most current adobe flash player version''', '''how many episodes are there in dragon ball z''', '''what is the first step in the evolution of the eye''', '''where is gall bladder situated in human body''', '''what is the main mineral in lithium batteries''', '''who is the president of usa right now''', '''where do the greasers live in the outsiders''', '''panda is a national animal of which country''', '''what is the name of manchester united stadium''', ] lowercase : str = tokenizer(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @slow def __lowerCamelCase ( self ): lowercase : int = RagTokenizer.from_pretrained('''facebook/rag-sequence-nq''' ) lowercase : Tuple = [ '''who got the first nobel prize in physics''', '''when is the next deadpool movie being released''', '''which mode is used for short wave broadcast service''', '''who is the owner of reading football club''', '''when is the next scandal episode coming out''', '''when is the last time the philadelphia won the superbowl''', '''what is the most current adobe flash player version''', '''how many episodes are there in dragon ball z''', '''what is the first step in the evolution of the eye''', '''where is gall bladder situated in human body''', '''what is the main mineral in lithium batteries''', '''who is the president of usa right now''', '''where do the greasers live in the outsiders''', '''panda is a national animal of which country''', '''what is the name of manchester united stadium''', ] lowercase : Any = tokenizer(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ )
173
0
import logging import os import sys from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import SeqaSeqTrainer from seqaseq_training_args import SeqaSeqTrainingArguments import transformers from transformers import ( AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer, HfArgumentParser, MBartTokenizer, MBartTokenizerFast, set_seed, ) from transformers.trainer_utils import EvaluationStrategy, is_main_process from transformers.training_args import ParallelMode from utils import ( SeqaSeqDataCollator, SeqaSeqDataset, assert_all_frozen, build_compute_metrics_fn, check_output_dir, freeze_embeds, freeze_params, lmap, save_json, use_task_specific_params, write_txt_file, ) a_ : Optional[Any] = logging.getLogger(__name__) @dataclass class _snake_case : _lowercase : str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) _lowercase : Optional[str] = field( default=A__ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=A__ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=A__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) _lowercase : bool = field(default=A__ , metadata={'''help''': '''Whether tp freeze the encoder.'''} ) _lowercase : bool = field(default=A__ , metadata={'''help''': '''Whether to freeze the embeddings.'''} ) @dataclass class _snake_case : _lowercase : str = field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''} ) _lowercase : Optional[str] = field( default='''summarization''' , metadata={'''help''': '''Task name, summarization (or summarization_{dataset} for pegasus) or translation'''} , ) _lowercase : Optional[int] = field( default=10_24 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) _lowercase : Optional[int] = field( default=1_28 , metadata={ '''help''': ( '''The maximum total sequence length for target text after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) _lowercase : Optional[int] = field( default=1_42 , metadata={ '''help''': ( '''The maximum total sequence length for validation target text after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded. ''' '''This argument is also used to override the ``max_length`` param of ``model.generate``, which is used ''' '''during ``evaluate`` and ``predict``.''' ) } , ) _lowercase : Optional[int] = field( default=1_42 , metadata={ '''help''': ( '''The maximum total sequence length for test target text after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) _lowercase : Optional[int] = field(default=-1 , metadata={'''help''': '''# training examples. -1 means use all.'''} ) _lowercase : Optional[int] = field(default=-1 , metadata={'''help''': '''# validation examples. -1 means use all.'''} ) _lowercase : Optional[int] = field(default=-1 , metadata={'''help''': '''# test examples. -1 means use all.'''} ) _lowercase : Optional[str] = field(default=A__ , metadata={'''help''': '''Source language id for translation.'''} ) _lowercase : Optional[str] = field(default=A__ , metadata={'''help''': '''Target language id for translation.'''} ) _lowercase : Optional[int] = field(default=A__ , metadata={'''help''': '''# num_beams to use for evaluation.'''} ) _lowercase : bool = field( default=A__ , metadata={'''help''': '''If only pad tokens should be ignored. This assumes that `config.pad_token_id` is defined.'''} , ) def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): logger.info(F'''***** {split} metrics *****''') for key in sorted(metrics.keys()): logger.info(F''' {key} = {metrics[key]}''') save_json(_UpperCAmelCase , os.path.join(_UpperCAmelCase , F'''{split}_results.json''')) def lowerCamelCase__ (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. SCREAMING_SNAKE_CASE = HfArgumentParser((ModelArguments, DataTrainingArguments, SeqaSeqTrainingArguments)) if len(sys.argv) == 2 and sys.argv[1].endswith('.json'): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1])) else: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = parser.parse_args_into_dataclasses() check_output_dir(_UpperCAmelCase) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.parallel_mode == ParallelMode.DISTRIBUTED) , training_args.fpaa , ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank): transformers.utils.logging.set_verbosity_info() logger.info('Training/evaluation parameters %s' , _UpperCAmelCase) # Set seed set_seed(training_args.seed) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) SCREAMING_SNAKE_CASE = ('encoder_layerdrop', 'decoder_layerdrop', 'dropout', 'attention_dropout') for p in extra_model_params: if getattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): assert hasattr(_UpperCAmelCase , _UpperCAmelCase), F'''({config.__class__.__name__}) doesn\'t have a `{p}` attribute''' setattr(_UpperCAmelCase , _UpperCAmelCase , getattr(_UpperCAmelCase , _UpperCAmelCase)) SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) SCREAMING_SNAKE_CASE = AutoModelForSeqaSeqLM.from_pretrained( model_args.model_name_or_path , from_tf='.ckpt' in model_args.model_name_or_path , config=_UpperCAmelCase , cache_dir=model_args.cache_dir , ) # use task specific params use_task_specific_params(_UpperCAmelCase , data_args.task) # set num_beams for evaluation if data_args.eval_beams is None: SCREAMING_SNAKE_CASE = model.config.num_beams # set decoder_start_token_id for MBart if model.config.decoder_start_token_id is None and isinstance(_UpperCAmelCase , (MBartTokenizer, MBartTokenizerFast)): assert ( data_args.tgt_lang is not None and data_args.src_lang is not None ), "mBart requires --tgt_lang and --src_lang" if isinstance(_UpperCAmelCase , _UpperCAmelCase): SCREAMING_SNAKE_CASE = tokenizer.lang_code_to_id[data_args.tgt_lang] else: SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(data_args.tgt_lang) if model_args.freeze_embeds: freeze_embeds(_UpperCAmelCase) if model_args.freeze_encoder: freeze_params(model.get_encoder()) assert_all_frozen(model.get_encoder()) SCREAMING_SNAKE_CASE = SeqaSeqDataset # Get datasets SCREAMING_SNAKE_CASE = ( dataset_class( _UpperCAmelCase , type_path='train' , data_dir=data_args.data_dir , n_obs=data_args.n_train , max_target_length=data_args.max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or '' , ) if training_args.do_train else None ) SCREAMING_SNAKE_CASE = ( dataset_class( _UpperCAmelCase , type_path='val' , data_dir=data_args.data_dir , n_obs=data_args.n_val , max_target_length=data_args.val_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or '' , ) if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO else None ) SCREAMING_SNAKE_CASE = ( dataset_class( _UpperCAmelCase , type_path='test' , data_dir=data_args.data_dir , n_obs=data_args.n_test , max_target_length=data_args.test_max_target_length , max_source_length=data_args.max_source_length , prefix=model.config.prefix or '' , ) if training_args.do_predict else None ) # Initialize our Trainer SCREAMING_SNAKE_CASE = ( build_compute_metrics_fn(data_args.task , _UpperCAmelCase) if training_args.predict_with_generate else None ) SCREAMING_SNAKE_CASE = SeqaSeqTrainer( model=_UpperCAmelCase , args=_UpperCAmelCase , data_args=_UpperCAmelCase , train_dataset=_UpperCAmelCase , eval_dataset=_UpperCAmelCase , data_collator=SeqaSeqDataCollator( _UpperCAmelCase , _UpperCAmelCase , model.config.decoder_start_token_id , training_args.tpu_num_cores) , compute_metrics=_UpperCAmelCase , tokenizer=_UpperCAmelCase , ) SCREAMING_SNAKE_CASE = {} # Training if training_args.do_train: logger.info('*** Train ***') SCREAMING_SNAKE_CASE = trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path) else None) SCREAMING_SNAKE_CASE = train_result.metrics SCREAMING_SNAKE_CASE = data_args.n_train trainer.save_model() # this also saves the tokenizer if trainer.is_world_process_zero(): handle_metrics('train' , _UpperCAmelCase , training_args.output_dir) all_metrics.update(_UpperCAmelCase) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir , 'trainer_state.json')) # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) tokenizer.save_pretrained(training_args.output_dir) # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***') SCREAMING_SNAKE_CASE = trainer.evaluate(metric_key_prefix='val') SCREAMING_SNAKE_CASE = data_args.n_val SCREAMING_SNAKE_CASE = round(metrics['val_loss'] , 4) if trainer.is_world_process_zero(): handle_metrics('val' , _UpperCAmelCase , training_args.output_dir) all_metrics.update(_UpperCAmelCase) if training_args.do_predict: logger.info('*** Predict ***') SCREAMING_SNAKE_CASE = trainer.predict(test_dataset=_UpperCAmelCase , metric_key_prefix='test') SCREAMING_SNAKE_CASE = test_output.metrics SCREAMING_SNAKE_CASE = data_args.n_test if trainer.is_world_process_zero(): SCREAMING_SNAKE_CASE = round(metrics['test_loss'] , 4) handle_metrics('test' , _UpperCAmelCase , training_args.output_dir) all_metrics.update(_UpperCAmelCase) if training_args.predict_with_generate: SCREAMING_SNAKE_CASE = tokenizer.batch_decode( test_output.predictions , skip_special_tokens=_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase) SCREAMING_SNAKE_CASE = lmap(str.strip , _UpperCAmelCase) write_txt_file(_UpperCAmelCase , os.path.join(training_args.output_dir , 'test_generations.txt')) if trainer.is_world_process_zero(): save_json(_UpperCAmelCase , os.path.join(training_args.output_dir , 'all_results.json')) return all_metrics def lowerCamelCase__ (_UpperCAmelCase): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
137
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy a_ : List[Any] = logging.getLogger(__name__) a_ : Dict = 'pytorch_model.bin' @dataclasses.dataclass class _snake_case : _lowercase : str = dataclasses.field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models.'''} ) _lowercase : Optional[str] = dataclasses.field( default=A__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co.'''} , ) @dataclasses.dataclass class _snake_case : _lowercase : str = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the training data.'''} ) _lowercase : str = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the data to predict on.'''} ) _lowercase : Optional[str] = dataclasses.field( default=A__ , metadata={'''help''': '''A csv or a json file containing the validation data.'''} ) _lowercase : Optional[str] = dataclasses.field( default=A__ , metadata={'''help''': '''The name of the task to train on.'''} , ) _lowercase : Optional[List[str]] = dataclasses.field( default=A__ , metadata={'''help''': '''The list of labels for the task.'''} ) @dataclasses.dataclass class _snake_case : _lowercase : str = dataclasses.field( metadata={'''help''': '''The output directory where the model predictions and checkpoints will be written.'''} ) _lowercase : Optional[str] = dataclasses.field( default='''accuracy''' , metadata={'''help''': '''The evaluation metric used for the task.'''} ) _lowercase : Optional[str] = dataclasses.field( default='''no''' , metadata={ '''help''': '''The evaluation strategy to adopt during training. Possible values are: ["no", "step", "epoch]''' } , ) _lowercase : Optional[int] = dataclasses.field( default=10 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) _lowercase : Optional[float] = dataclasses.field( default=0.0 , metadata={ '''help''': '''How much the specified evaluation metric must improve to satisfy early stopping conditions.''' } , ) _lowercase : Optional[bool] = dataclasses.field( default=A__ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the confidence score.'''} , ) _lowercase : Optional[bool] = dataclasses.field( default=A__ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the validation performance.'''} , ) _lowercase : Optional[bool] = dataclasses.field( default=A__ , metadata={'''help''': '''Whether to fine-tune on labeled data after pseudo training.'''} , ) _lowercase : Optional[float] = dataclasses.field( default=0.0 , metadata={'''help''': '''Confidence threshold for pseudo-labeled data filtering.'''} , ) _lowercase : Optional[int] = dataclasses.field( default=1_00 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) _lowercase : Optional[int] = dataclasses.field( default=A__ , metadata={'''help''': '''Random seed for initialization.'''} , ) def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): SCREAMING_SNAKE_CASE = datasets.concatenate_datasets([infer_input, infer_output] , axis=1) if args.do_filter_by_confidence: SCREAMING_SNAKE_CASE = dataset.filter(lambda _UpperCAmelCase: example["probability"] > args.confidence_threshold) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 SCREAMING_SNAKE_CASE = int(eval_result * len(_UpperCAmelCase)) print(_UpperCAmelCase) SCREAMING_SNAKE_CASE = dataset.sort('probability' , reverse=_UpperCAmelCase) SCREAMING_SNAKE_CASE = dataset.select(range(_UpperCAmelCase)) SCREAMING_SNAKE_CASE = dataset.remove_columns(['label', 'probability']) SCREAMING_SNAKE_CASE = dataset.rename_column('prediction' , 'label') SCREAMING_SNAKE_CASE = dataset.map(lambda _UpperCAmelCase: {"label": idalabel[example["label"]]}) SCREAMING_SNAKE_CASE = dataset.shuffle(seed=args.seed) SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , F'''train_pseudo.{args.data_file_extension}''') if args.data_file_extension == "csv": dataset.to_csv(_UpperCAmelCase , index=_UpperCAmelCase) else: dataset.to_json(_UpperCAmelCase) def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase): SCREAMING_SNAKE_CASE = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO , ) logger.info(accelerator.state) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() SCREAMING_SNAKE_CASE = STModelArguments(model_name_or_path=_UpperCAmelCase) SCREAMING_SNAKE_CASE = STDataArguments(train_file=_UpperCAmelCase , infer_file=_UpperCAmelCase) SCREAMING_SNAKE_CASE = STTrainingArguments(output_dir=_UpperCAmelCase) SCREAMING_SNAKE_CASE = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(_UpperCAmelCase).items(): setattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase) for key, value in kwargs.items(): if hasattr(_UpperCAmelCase , _UpperCAmelCase): setattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase) # Sanity checks SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None SCREAMING_SNAKE_CASE = args.train_file SCREAMING_SNAKE_CASE = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None SCREAMING_SNAKE_CASE = args.eval_file for key in data_files: SCREAMING_SNAKE_CASE = data_files[key].split('.')[-1] assert extension in ["csv", "json"], F'''`{key}_file` should be a csv or a json file.''' if args.data_file_extension is None: SCREAMING_SNAKE_CASE = extension else: assert extension == args.data_file_extension, F'''`{key}_file` should be a {args.data_file_extension} file`.''' assert ( args.eval_metric in datasets.list_metrics() ), F'''{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.''' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) logger.info('Creating the initial data directory for self-training...') SCREAMING_SNAKE_CASE = F'''{args.output_dir}/self-train_iter-{{}}'''.format SCREAMING_SNAKE_CASE = data_dir_format(0) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=_UpperCAmelCase) os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase) accelerator.wait_for_everyone() SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = False # Show the progress bar SCREAMING_SNAKE_CASE = tqdm(range(args.max_selftrain_iterations) , disable=not accelerator.is_local_main_process) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations)): SCREAMING_SNAKE_CASE = data_dir_format(_UpperCAmelCase) assert os.path.exists(_UpperCAmelCase) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'stage-1') SCREAMING_SNAKE_CASE = { 'accelerator': accelerator, 'model_name_or_path': args.model_name_or_path, 'cache_dir': args.cache_dir, 'do_train': True, 'train_file': data_files['train'] if iteration == 0 else data_files['train_pseudo'], 'do_eval': True if args.eval_file is not None else False, 'eval_file': data_files['eval'], 'do_predict': True, 'infer_file': data_files['infer'], 'task_name': args.task_name, 'label_list': args.label_list, 'output_dir': current_output_dir, 'eval_metric': args.eval_metric, 'evaluation_strategy': args.evaluation_strategy, 'early_stopping_patience': args.early_stopping_patience, 'early_stopping_threshold': args.early_stopping_threshold, 'seed': args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(_UpperCAmelCase , _UpperCAmelCase): arguments_dict.update({key: value}) SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'best-checkpoint' , _UpperCAmelCase) if os.path.exists(_UpperCAmelCase): logger.info( 'Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.' , _UpperCAmelCase , _UpperCAmelCase , ) else: logger.info('***** Running self-training: iteration: %d, stage: 1 *****' , _UpperCAmelCase) finetune(**_UpperCAmelCase) accelerator.wait_for_everyone() assert os.path.exists(_UpperCAmelCase) logger.info('Self-training job completed: iteration: %d, stage: 1.' , _UpperCAmelCase) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'best-checkpoint') SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'stage-2') # Update arguments_dict SCREAMING_SNAKE_CASE = model_path SCREAMING_SNAKE_CASE = data_files['train'] SCREAMING_SNAKE_CASE = current_output_dir SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'best-checkpoint' , _UpperCAmelCase) if os.path.exists(_UpperCAmelCase): logger.info( 'Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.' , _UpperCAmelCase , _UpperCAmelCase , ) else: logger.info('***** Running self-training: iteration: %d, stage: 2 *****' , _UpperCAmelCase) finetune(**_UpperCAmelCase) accelerator.wait_for_everyone() assert os.path.exists(_UpperCAmelCase) logger.info('Self-training job completed: iteration: %d, stage: 2.' , _UpperCAmelCase) SCREAMING_SNAKE_CASE = iteration SCREAMING_SNAKE_CASE = data_dir_format(iteration + 1) SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(os.path.join(_UpperCAmelCase , 'best-checkpoint')) SCREAMING_SNAKE_CASE = config.idalabel SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'eval_results_best-checkpoint.json') SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'test_results_best-checkpoint.json') assert os.path.exists(_UpperCAmelCase) with open(_UpperCAmelCase , 'r') as f: SCREAMING_SNAKE_CASE = float(json.load(_UpperCAmelCase)[args.eval_metric]) SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , 'infer_output_best-checkpoint.csv') assert os.path.exists(_UpperCAmelCase) # Loading the dataset from local csv or json files. SCREAMING_SNAKE_CASE = load_dataset(args.data_file_extension , data_files={'data': data_files['infer']})['data'] SCREAMING_SNAKE_CASE = load_dataset('csv' , data_files={'data': infer_output_file})['data'] if accelerator.is_main_process: os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase) shutil.copy(_UpperCAmelCase , os.path.join(_UpperCAmelCase , F'''eval_results_iter-{iteration}.json''')) if os.path.exists(_UpperCAmelCase): shutil.copy(_UpperCAmelCase , os.path.join(_UpperCAmelCase , F'''test_results_iter-{iteration}.json''')) create_pseudo_labeled_data(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase) accelerator.wait_for_everyone() SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , F'''train_pseudo.{args.data_file_extension}''') if args.evaluation_strategy != IntervalStrategy.NO.value: SCREAMING_SNAKE_CASE = eval_result if best_iteration is None: SCREAMING_SNAKE_CASE = new_iteration SCREAMING_SNAKE_CASE = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: SCREAMING_SNAKE_CASE = new_iteration SCREAMING_SNAKE_CASE = new_eval_result SCREAMING_SNAKE_CASE = 0 else: if new_eval_result == best_eval_result: SCREAMING_SNAKE_CASE = new_iteration SCREAMING_SNAKE_CASE = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: SCREAMING_SNAKE_CASE = True progress_bar.update(1) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info('Best iteration: %d' , _UpperCAmelCase) logger.info('Best evaluation result: %s = %f' , args.eval_metric , _UpperCAmelCase) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_UpperCAmelCase , F'''eval_results_iter-{iteration}.json''') , os.path.join(_UpperCAmelCase , 'eval_results_best-iteration.json') , ) else: # Assume that the last iteration is the best logger.info('Best iteration: %d' , args.max_selftrain_iterations - 1) logger.info('Best evaluation result: %s = %f' , args.eval_metric , _UpperCAmelCase) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_UpperCAmelCase , F'''eval_results_iter-{args.max_selftrain_iterations - 1}.json''') , os.path.join(_UpperCAmelCase , 'eval_results_best-iteration.json') , )
137
1
'''simple docstring''' import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) set_seed(770) _snake_case = { 'c_attn': 'att_proj', 'c_proj': 'out_proj', 'c_fc': 'in_proj', 'transformer.': '', 'h.': 'layers.', 'ln_1': 'layernorm_1', 'ln_2': 'layernorm_2', 'ln_f': 'layernorm_final', 'wpe': 'position_embeds_layer', 'wte': 'input_embeds_layer', } _snake_case = { 'text_small': { 'repo_id': 'suno/bark', 'file_name': 'text.pt', }, 'coarse_small': { 'repo_id': 'suno/bark', 'file_name': 'coarse.pt', }, 'fine_small': { 'repo_id': 'suno/bark', 'file_name': 'fine.pt', }, 'text': { 'repo_id': 'suno/bark', 'file_name': 'text_2.pt', }, 'coarse': { 'repo_id': 'suno/bark', 'file_name': 'coarse_2.pt', }, 'fine': { 'repo_id': 'suno/bark', 'file_name': 'fine_2.pt', }, } _snake_case = os.path.dirname(os.path.abspath(__file__)) _snake_case = os.path.join(os.path.expanduser('~'), '.cache') _snake_case = os.path.join(os.getenv('XDG_CACHE_HOME', default_cache_dir), 'suno', 'bark_v0') def _A ( snake_case , snake_case=False ) -> Optional[int]: _lowercase : List[Any] = model_type if use_small: key += "_small" return os.path.join(_lowerCAmelCase , REMOTE_MODEL_PATHS[key]["file_name"] ) def _A ( snake_case , snake_case ) -> Dict: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) hf_hub_download(repo_id=_lowerCAmelCase , filename=_lowerCAmelCase , local_dir=_lowerCAmelCase ) def _A ( snake_case , snake_case , snake_case=False , snake_case="text" ) -> Dict: if model_type == "text": _lowercase : Tuple = BarkSemanticModel _lowercase : str = BarkSemanticConfig _lowercase : Optional[int] = BarkSemanticGenerationConfig elif model_type == "coarse": _lowercase : Tuple = BarkCoarseModel _lowercase : int = BarkCoarseConfig _lowercase : List[Any] = BarkCoarseGenerationConfig elif model_type == "fine": _lowercase : List[Any] = BarkFineModel _lowercase : Optional[Any] = BarkFineConfig _lowercase : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() _lowercase : Optional[Any] = F'''{model_type}_small''' if use_small else model_type _lowercase : Tuple = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(_lowerCAmelCase ): logger.info(F'''{model_type} model not found, downloading into `{CACHE_DIR}`.''' ) _download(model_info["repo_id"] , model_info["file_name"] ) _lowercase : Dict = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase ) # this is a hack _lowercase : int = checkpoint["""model_args"""] if "input_vocab_size" not in model_args: _lowercase : str = model_args["""vocab_size"""] _lowercase : Any = model_args["""vocab_size"""] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments _lowercase : Union[str, Any] = model_args.pop("n_head" ) _lowercase : Any = model_args.pop("n_embd" ) _lowercase : Union[str, Any] = model_args.pop("n_layer" ) _lowercase : Union[str, Any] = ConfigClass(**checkpoint["model_args"] ) _lowercase : Tuple = ModelClass(config=_lowerCAmelCase ) _lowercase : str = GenerationConfigClass() _lowercase : Tuple = model_generation_config _lowercase : Dict = checkpoint["""model"""] # fixup checkpoint _lowercase : Optional[Any] = """_orig_mod.""" for k, v in list(state_dict.items() ): if k.startswith(_lowerCAmelCase ): # replace part of the key with corresponding layer name in HF implementation _lowercase : Optional[int] = k[len(_lowerCAmelCase ) :] for old_layer_name in new_layer_name_dict: _lowercase : Tuple = new_k.replace(_lowerCAmelCase , new_layer_name_dict[old_layer_name] ) _lowercase : List[Any] = state_dict.pop(_lowerCAmelCase ) _lowercase : Union[str, Any] = set(state_dict.keys() ) - set(model.state_dict().keys() ) _lowercase : Union[str, Any] = {k for k in extra_keys if not k.endswith(".attn.bias" )} _lowercase : Optional[Any] = set(model.state_dict().keys() ) - set(state_dict.keys() ) _lowercase : str = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(_lowerCAmelCase ) != 0: raise ValueError(F'''extra keys found: {extra_keys}''' ) if len(_lowerCAmelCase ) != 0: raise ValueError(F'''missing keys: {missing_keys}''' ) model.load_state_dict(_lowerCAmelCase , strict=_lowerCAmelCase ) _lowercase : List[Any] = model.num_parameters(exclude_embeddings=_lowerCAmelCase ) _lowercase : Any = checkpoint["""best_val_loss"""].item() logger.info(F'''model loaded: {round(n_params/1E6 , 1 )}M params, {round(_lowerCAmelCase , 3 )} loss''' ) model.eval() model.to(_lowerCAmelCase ) del checkpoint, state_dict return model def _A ( snake_case , snake_case=False , snake_case="text" ) -> Optional[int]: if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() _lowercase : str = """cpu""" # do conversion on cpu _lowercase : Union[str, Any] = _get_ckpt_path(_lowerCAmelCase , use_small=_lowerCAmelCase ) _lowercase : Union[str, Any] = _load_model(_lowerCAmelCase , _lowerCAmelCase , model_type=_lowerCAmelCase , use_small=_lowerCAmelCase ) # load bark initial model _lowercase : Optional[int] = _bark_load_model(_lowerCAmelCase , "cpu" , model_type=_lowerCAmelCase , use_small=_lowerCAmelCase ) if model_type == "text": _lowercase : int = bark_model["""model"""] if model.num_parameters(exclude_embeddings=_lowerCAmelCase ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model _lowercase : Tuple = 5 _lowercase : Union[str, Any] = 10 if model_type in ["text", "coarse"]: _lowercase : Optional[int] = torch.randint(2_56 , (batch_size, sequence_length) , dtype=torch.int ) _lowercase : Optional[int] = bark_model(_lowerCAmelCase )[0] _lowercase : Any = model(_lowerCAmelCase ) # take last logits _lowercase : Optional[int] = output_new_model_total.logits[:, [-1], :] else: _lowercase : str = 3 _lowercase : Union[str, Any] = 8 _lowercase : Any = torch.randint(2_56 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) _lowercase : int = model(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Optional[Any] = bark_model(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Tuple = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1E-3: raise ValueError("initial and new outputs are not equal" ) Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) def _A ( snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ) -> List[str]: _lowercase : List[str] = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : List[str] = BarkSemanticConfig.from_pretrained(os.path.join(_lowerCAmelCase , "config.json" ) ) _lowercase : List[Any] = BarkCoarseConfig.from_pretrained(os.path.join(_lowerCAmelCase , "config.json" ) ) _lowercase : Any = BarkFineConfig.from_pretrained(os.path.join(_lowerCAmelCase , "config.json" ) ) _lowercase : int = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) _lowercase : int = BarkSemanticModel.from_pretrained(_lowerCAmelCase ) _lowercase : str = BarkCoarseModel.from_pretrained(_lowerCAmelCase ) _lowercase : Tuple = BarkFineModel.from_pretrained(_lowerCAmelCase ) _lowercase : int = EncodecModel.from_pretrained("facebook/encodec_24khz" ) _lowercase : List[Any] = BarkConfig.from_sub_model_configs( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) _lowercase : int = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) _lowercase : Tuple = BarkModel(_lowerCAmelCase ) _lowercase : Optional[Any] = semantic _lowercase : List[Any] = coarseAcoustic _lowercase : List[str] = fineAcoustic _lowercase : List[str] = codec _lowercase : Dict = bark_generation_config Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) bark.save_pretrained(_lowerCAmelCase , repo_id=_lowerCAmelCase , push_to_hub=_lowerCAmelCase ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument('model_type', type=str, help='text, coarse or fine.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--is_small', action='store_true', help='convert the small version instead of the large.') _snake_case = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
371
'''simple docstring''' from timeit import timeit def _A ( snake_case ) -> int: if number < 0: raise ValueError("the value of input must not be negative" ) _lowercase : Union[str, Any] = 0 while number: number &= number - 1 result += 1 return result def _A ( snake_case ) -> int: if number < 0: raise ValueError("the value of input must not be negative" ) _lowercase : int = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def _A ( ) -> None: def do_benchmark(snake_case ) -> None: _lowercase : Optional[int] = "import __main__ as z" print(F'''Benchmark when {number = }:''' ) print(F'''{get_set_bits_count_using_modulo_operator(snake_case ) = }''' ) _lowercase : int = timeit("z.get_set_bits_count_using_modulo_operator(25)" , setup=snake_case ) print(F'''timeit() runs in {timing} seconds''' ) print(F'''{get_set_bits_count_using_brian_kernighans_algorithm(snake_case ) = }''' ) _lowercase : Optional[int] = timeit( "z.get_set_bits_count_using_brian_kernighans_algorithm(25)" , setup=snake_case , ) print(F'''timeit() runs in {timing} seconds''' ) for number in (25, 37, 58, 0): do_benchmark(snake_case ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
199
0
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def _A ( SCREAMING_SNAKE_CASE : int = 3 ): """simple docstring""" if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise TypeError("number of qubits must be a integer." ) if number_of_qubits <= 0: raise ValueError("number of qubits must be > 0." ) if math.floor(SCREAMING_SNAKE_CASE ) != number_of_qubits: raise ValueError("number of qubits must be exact integer." ) if number_of_qubits > 10: raise ValueError("number of qubits too large to simulate(>10)." ) a__ : Optional[int] =QuantumRegister(SCREAMING_SNAKE_CASE , "qr" ) a__ : List[Any] =ClassicalRegister(SCREAMING_SNAKE_CASE , "cr" ) a__ : int =QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) a__ : Optional[Any] =number_of_qubits for i in range(SCREAMING_SNAKE_CASE ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(SCREAMING_SNAKE_CASE ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(SCREAMING_SNAKE_CASE , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # simulate with 10000 shots a__ : Optional[Any] =Aer.get_backend("qasm_simulator" ) a__ : List[str] =execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=10_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print( F"""Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}""" )
95
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def lowerCamelCase__ ( _A , _A ): if inductance <= 0: raise ValueError('Inductance cannot be 0 or negative' ) elif capacitance <= 0: raise ValueError('Capacitance cannot be 0 or negative' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
297
0
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = 1_0 def A__ ( self): lowercase = [1, 2, 3, 4] lowercase = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(A__ ,self.block_size ,0) ,A__) def A__ ( self): lowercase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] lowercase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(A__ ,self.block_size ,0) ,A__) def A__ ( self): lowercase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0, 1_1, 1_2, 1_3] lowercase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(A__ ,self.block_size ,0) ,A__) def A__ ( self): lowercase = '''It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.''' lowercase , lowercase = process_story(A__) self.assertEqual(A__ ,[]) def A__ ( self): lowercase = '''''' lowercase , lowercase = process_story(A__) self.assertEqual(A__ ,[]) self.assertEqual(A__ ,[]) def A__ ( self): lowercase = ( '''It was the year of Our Lord one thousand seven hundred and ''' '''seventy-five\n\nSpiritual revelations were conceded to England ''' '''at that favoured period, as at this.\n@highlight\n\nIt was the best of times''' ) lowercase , lowercase = process_story(A__) lowercase = [ '''It was the year of Our Lord one thousand seven hundred and seventy-five.''', '''Spiritual revelations were conceded to England at that favoured period, as at this.''', ] self.assertEqual(A__ ,A__) lowercase = ['''It was the best of times.'''] self.assertEqual(A__ ,A__) def A__ ( self): lowercase = torch.tensor([1, 2, 3, 4]) lowercase = torch.tensor([1, 1, 1, 1]) np.testing.assert_array_equal(build_mask(A__ ,0).numpy() ,expected.numpy()) def A__ ( self): lowercase = torch.tensor([1, 2, 3, 4, 2_3, 2_3, 2_3]) lowercase = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(A__ ,2_3).numpy() ,expected.numpy()) def A__ ( self): lowercase = torch.tensor([8, 2, 3, 4, 1, 1, 1]) lowercase = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(A__ ,1).numpy() ,expected.numpy()) def A__ ( self): lowercase = 1_0_1 lowercase = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 1_0_1, 5, 6], [1, 1_0_1, 3, 4, 1_0_1, 6]]) lowercase = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]]) lowercase = compute_token_type_ids(A__ ,A__) np.testing.assert_array_equal(A__ ,A__)
97
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Dict = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowercase__ :Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
97
1
"""simple docstring""" import argparse import os import re import packaging.version lowerCamelCase_ : Any = """examples/""" lowerCamelCase_ : Optional[Any] = { """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } lowerCamelCase_ : Any = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } lowerCamelCase_ : Optional[Any] = """README.md""" def _A ( lowercase , lowercase , lowercase ): """simple docstring""" with open(_lowercase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: a =f.read() a =REPLACE_PATTERNS[pattern] a =replace.replace('''VERSION''' , _lowercase ) a =re_pattern.sub(_lowercase , _lowercase ) with open(_lowercase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowercase ) def _A ( lowercase ): """simple docstring""" for folder, directories, fnames in os.walk(_lowercase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowercase , _lowercase ) , _lowercase , pattern='''examples''' ) def _A ( lowercase , lowercase=False ): """simple docstring""" for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowercase , _lowercase , _lowercase ) if not patch: update_version_in_examples(_lowercase ) def _A ( ): """simple docstring""" a ="""🤗 Transformers currently provides the following architectures""" a ="""1. Want to contribute a new model?""" with open(_lowercase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: a =f.readlines() # Find the start of the list. a =0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 a =start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): a =lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(_lowercase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowercase ) def _A ( ): """simple docstring""" with open(REPLACE_FILES['''init'''] , '''r''' ) as f: a =f.read() a =REPLACE_PATTERNS["""init"""][0].search(_lowercase ).groups()[0] return packaging.version.parse(_lowercase ) def _A ( lowercase=False ): """simple docstring""" a =get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: a =default_version.base_version elif patch: a =f'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: a =f'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. a =input(f'''Which version are you releasing? [{default_version}]''' ) if len(_lowercase ) == 0: a =default_version print(f'''Updating version to {version}.''' ) global_version_update(_lowercase , patch=_lowercase ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def _A ( ): """simple docstring""" a =get_version() a =f'''{current_version.major}.{current_version.minor + 1}.0.dev0''' a =current_version.base_version # Check with the user we got that right. a =input(f'''Which version are we developing now? [{dev_version}]''' ) if len(_lowercase ) == 0: a =dev_version print(f'''Updating version to {version}.''' ) global_version_update(_lowercase ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": lowerCamelCase_ : List[str] = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") lowerCamelCase_ : List[str] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
81
"""simple docstring""" import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __a = logging.getLogger(__name__) class lowerCamelCase ( _lowerCAmelCase ): '''simple docstring''' _A : Union[str, Any] = """token-classification""" def __init__( self: Any , snake_case: Tuple ) -> List[Any]: if type(snake_case ) == dict: snake_case_ :Optional[int] = Namespace(**snake_case ) snake_case_ :Optional[int] = import_module("""tasks""" ) try: snake_case_ :Any = getattr(snake_case , hparams.task_type ) snake_case_ :TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) snake_case_ :Any = self.token_classification_task.get_labels(hparams.labels ) snake_case_ :str = CrossEntropyLoss().ignore_index super().__init__(snake_case , len(self.labels ) , self.mode ) def lowerCAmelCase_ ( self: Dict , **snake_case: List[Any] ) -> Any: return self.model(**snake_case ) def lowerCAmelCase_ ( self: str , snake_case: Tuple , snake_case: List[Any] ) -> Optional[int]: snake_case_ :List[str] = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": snake_case_ :List[str] = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids snake_case_ :Optional[Any] = self(**snake_case ) snake_case_ :List[str] = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def lowerCAmelCase_ ( self: int ) -> Dict: snake_case_ :List[Any] = self.hparams for mode in ["train", "dev", "test"]: snake_case_ :Optional[int] = self._feature_file(snake_case ) if os.path.exists(snake_case ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , snake_case ) snake_case_ :Optional[int] = torch.load(snake_case ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) snake_case_ :Optional[int] = self.token_classification_task.read_examples_from_file(args.data_dir , snake_case ) snake_case_ :Any = self.token_classification_task.convert_examples_to_features( snake_case , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["""xlnet"""] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["""xlnet"""] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=snake_case , pad_on_left=bool(self.config.model_type in ["""xlnet"""] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("""Saving features into cached file %s""" , snake_case ) torch.save(snake_case , snake_case ) def lowerCAmelCase_ ( self: Optional[int] , snake_case: int , snake_case: int , snake_case: bool = False ) -> DataLoader: snake_case_ :int = self._feature_file(snake_case ) logger.info("""Loading features from cached file %s""" , snake_case ) snake_case_ :str = torch.load(snake_case ) snake_case_ :Dict = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) snake_case_ :str = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: snake_case_ :List[Any] = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: snake_case_ :List[str] = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) snake_case_ :Any = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(snake_case , snake_case , snake_case , snake_case ) , batch_size=snake_case ) def lowerCAmelCase_ ( self: List[str] , snake_case: Dict , snake_case: Union[str, Any] ) -> List[str]: """Compute validation""" "" snake_case_ :List[str] = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": snake_case_ :Dict = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids snake_case_ :Dict = self(**snake_case ) snake_case_, snake_case_ :Dict = outputs[:2] snake_case_ :Union[str, Any] = logits.detach().cpu().numpy() snake_case_ :List[Any] = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def lowerCAmelCase_ ( self: List[Any] , snake_case: int ) -> Tuple: snake_case_ :Union[str, Any] = torch.stack([x["""val_loss"""] for x in outputs] ).mean() snake_case_ :Tuple = np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) snake_case_ :Tuple = np.argmax(snake_case , axis=2 ) snake_case_ :List[str] = np.concatenate([x["""target"""] for x in outputs] , axis=0 ) snake_case_ :Optional[Any] = dict(enumerate(self.labels ) ) snake_case_ :Dict = [[] for _ in range(out_label_ids.shape[0] )] snake_case_ :Dict = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) snake_case_ :str = { """val_loss""": val_loss_mean, """accuracy_score""": accuracy_score(snake_case , snake_case ), """precision""": precision_score(snake_case , snake_case ), """recall""": recall_score(snake_case , snake_case ), """f1""": fa_score(snake_case , snake_case ), } snake_case_ :List[Any] = dict(results.items() ) snake_case_ :Union[str, Any] = results return ret, preds_list, out_label_list def lowerCAmelCase_ ( self: Optional[Any] , snake_case: Dict ) -> Optional[Any]: # when stable snake_case_, snake_case_, snake_case_ :Tuple = self._eval_end(snake_case ) snake_case_ :str = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def lowerCAmelCase_ ( self: Tuple , snake_case: Optional[int] ) -> Any: # updating to test_epoch_end instead of deprecated test_end snake_case_, snake_case_, snake_case_ :Any = self._eval_end(snake_case ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 snake_case_ :Optional[int] = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def lowerCAmelCase_ ( snake_case: Any , snake_case: int ) -> Dict: # Add NER specific options BaseTransformer.add_model_specific_args(snake_case , snake_case ) parser.add_argument( """--task_type""" , default="""NER""" , type=snake_case , help="""Task type to fine tune in training (e.g. NER, POS, etc)""" ) parser.add_argument( """--max_seq_length""" , default=128 , type=snake_case , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--labels""" , default="""""" , type=snake_case , help="""Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.""" , ) parser.add_argument( """--gpus""" , default=0 , type=snake_case , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser if __name__ == "__main__": __a = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __a = NERTransformer.add_model_specific_args(parser, os.getcwd()) __a = parser.parse_args() __a = NERTransformer(args) __a = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __a = sorted(glob.glob(os.path.join(args.output_dir, "checkpoint-epoch=*.ckpt"), recursive=True)) __a = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
66
0
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class lowercase : def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=2 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , snake_case=1000 , ): snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope snake_case_ = range_bbox def a ( self ): snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment snake_case_ = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: snake_case_ = bbox[i, j, 3] snake_case_ = bbox[i, j, 1] snake_case_ = t if bbox[i, j, 2] < bbox[i, j, 0]: snake_case_ = bbox[i, j, 2] snake_case_ = bbox[i, j, 0] snake_case_ = t snake_case_ = tf.convert_to_tensor(snake_case ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): snake_case_ = TFLayoutLMModel(config=snake_case ) snake_case_ = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) snake_case_ = model(snake_case , snake_case , token_type_ids=snake_case ) snake_case_ = model(snake_case , snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): snake_case_ = TFLayoutLMForMaskedLM(config=snake_case ) snake_case_ = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): snake_case_ = self.num_labels snake_case_ = TFLayoutLMForSequenceClassification(config=snake_case ) snake_case_ = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): snake_case_ = self.num_labels snake_case_ = TFLayoutLMForTokenClassification(config=snake_case ) snake_case_ = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): snake_case_ = TFLayoutLMForQuestionAnswering(config=snake_case ) snake_case_ = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a ( self ): snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_tf class lowercase ( lowercase_ , lowercase_ , unittest.TestCase ): __SCREAMING_SNAKE_CASE : Dict = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __SCREAMING_SNAKE_CASE : Tuple = ( { '''feature-extraction''': TFLayoutLMModel, '''fill-mask''': TFLayoutLMForMaskedLM, '''text-classification''': TFLayoutLMForSequenceClassification, '''token-classification''': TFLayoutLMForTokenClassification, '''zero-shot''': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : List[Any] = True __SCREAMING_SNAKE_CASE : Any = 10 def a ( self ): snake_case_ = TFLayoutLMModelTester(self ) snake_case_ = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def a ( self ): self.config_tester.run_common_tests() def a ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def a ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def a ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def a ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) def a ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) @slow def a ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = TFLayoutLMModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip('Onnx compliancy broke with TF 2.10' ) def a ( self ): pass def __lowerCamelCase ( ): '''simple docstring''' snake_case_ = tf.convert_to_tensor([[101,1019,1014,1016,1037,12849,4747,1004,14246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,11300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,19274,2772,6205,27814,16147,16147,4343,2047,10283,10969,14389,1012,2338,102]] ) # noqa: E231 snake_case_ = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 snake_case_ = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 snake_case_ = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) snake_case_ = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class lowercase ( unittest.TestCase ): @slow def a ( self ): snake_case_ = TFLayoutLMModel.from_pretrained('microsoft/layoutlm-base-uncased' ) snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = prepare_layoutlm_batch_inputs() # forward pass snake_case_ = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the sequence output on [0, :3, :3] snake_case_ = tf.convert_to_tensor( [[0.17_85, -0.19_47, -0.04_25], [-0.32_54, -0.28_07, 0.25_53], [-0.53_91, -0.33_22, 0.33_64]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1e-3 ) ) # test the pooled output on [1, :3] snake_case_ = tf.convert_to_tensor([-0.65_80, -0.02_14, 0.85_52] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , snake_case , atol=1e-3 ) ) @slow def a ( self ): # initialize model with randomly initialized sequence classification head snake_case_ = TFLayoutLMForSequenceClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=2 ) snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = prepare_layoutlm_batch_inputs() # forward pass snake_case_ = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar snake_case_ = outputs.loss snake_case_ = (2,) self.assertEqual(loss.shape , snake_case ) # test the shape of the logits snake_case_ = outputs.logits snake_case_ = (2, 2) self.assertEqual(logits.shape , snake_case ) @slow def a ( self ): # initialize model with randomly initialized token classification head snake_case_ = TFLayoutLMForTokenClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=13 ) snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = prepare_layoutlm_batch_inputs() # forward pass snake_case_ = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) # test the shape of the logits snake_case_ = outputs.logits snake_case_ = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , snake_case ) @slow def a ( self ): # initialize model with randomly initialized token classification head snake_case_ = TFLayoutLMForQuestionAnswering.from_pretrained('microsoft/layoutlm-base-uncased' ) snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = prepare_layoutlm_batch_inputs() # forward pass snake_case_ = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the shape of the logits snake_case_ = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , snake_case ) self.assertEqual(outputs.end_logits.shape , snake_case )
200
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase ( unittest.TestCase ): def a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'bird' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) snake_case_ = pipe.prepare_image_inputs([canny_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'Chef in the kitchen' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) snake_case_ = pipe.prepare_image_inputs([pose_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
200
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging A__ : List[Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class lowercase__ ( snake_case__ ): _UpperCAmelCase :str = ["pixel_values"] def __init__( self : Optional[Any] , snake_case__ : bool = True , snake_case__ : Dict[str, int] = None , snake_case__ : PILImageResampling = PILImageResampling.BICUBIC , snake_case__ : bool = True , snake_case__ : Dict[str, int] = None , snake_case__ : bool = True , snake_case__ : Union[int, float] = 1 / 255 , snake_case__ : bool = True , snake_case__ : Optional[Union[float, List[float]]] = None , snake_case__ : Optional[Union[float, List[float]]] = None , snake_case__ : bool = True , **snake_case__ : Optional[int] , ): super().__init__(**snake_case__ ) lowerCamelCase_ : int =size if size is not None else {'''shortest_edge''': 224} lowerCamelCase_ : Optional[Any] =get_size_dict(snake_case__ , default_to_square=snake_case__ ) lowerCamelCase_ : str =crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} lowerCamelCase_ : Any =get_size_dict(snake_case__ , default_to_square=snake_case__ , param_name="crop_size" ) lowerCamelCase_ : Union[str, Any] =do_resize lowerCamelCase_ : List[Any] =size lowerCamelCase_ : Tuple =resample lowerCamelCase_ : Optional[Any] =do_center_crop lowerCamelCase_ : List[Any] =crop_size lowerCamelCase_ : Tuple =do_rescale lowerCamelCase_ : Optional[int] =rescale_factor lowerCamelCase_ : Tuple =do_normalize lowerCamelCase_ : List[Any] =image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowerCamelCase_ : Any =image_std if image_std is not None else OPENAI_CLIP_STD lowerCamelCase_ : Dict =do_convert_rgb def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : np.ndarray , snake_case__ : Dict[str, int] , snake_case__ : PILImageResampling = PILImageResampling.BICUBIC , snake_case__ : Optional[Union[str, ChannelDimension]] = None , **snake_case__ : Optional[Any] , ): lowerCamelCase_ : int =get_size_dict(snake_case__ , default_to_square=snake_case__ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) lowerCamelCase_ : int =get_resize_output_image_size(snake_case__ , size=size["shortest_edge"] , default_to_square=snake_case__ ) return resize(snake_case__ , size=snake_case__ , resample=snake_case__ , data_format=snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : Tuple , snake_case__ : np.ndarray , snake_case__ : Dict[str, int] , snake_case__ : Optional[Union[str, ChannelDimension]] = None , **snake_case__ : int , ): lowerCamelCase_ : int =get_size_dict(snake_case__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(snake_case__ , size=(size["height"], size["width"]) , data_format=snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : Any , snake_case__ : np.ndarray , snake_case__ : Union[int, float] , snake_case__ : Optional[Union[str, ChannelDimension]] = None , **snake_case__ : Any , ): return rescale(snake_case__ , scale=snake_case__ , data_format=snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : Any , snake_case__ : np.ndarray , snake_case__ : Union[float, List[float]] , snake_case__ : Union[float, List[float]] , snake_case__ : Optional[Union[str, ChannelDimension]] = None , **snake_case__ : List[str] , ): return normalize(snake_case__ , mean=snake_case__ , std=snake_case__ , data_format=snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : Tuple , snake_case__ : ImageInput , snake_case__ : bool = None , snake_case__ : Dict[str, int] = None , snake_case__ : PILImageResampling = None , snake_case__ : bool = None , snake_case__ : int = None , snake_case__ : bool = None , snake_case__ : float = None , snake_case__ : bool = None , snake_case__ : Optional[Union[float, List[float]]] = None , snake_case__ : Optional[Union[float, List[float]]] = None , snake_case__ : bool = None , snake_case__ : Optional[Union[str, TensorType]] = None , snake_case__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **snake_case__ : List[Any] , ): lowerCamelCase_ : Tuple =do_resize if do_resize is not None else self.do_resize lowerCamelCase_ : Union[str, Any] =size if size is not None else self.size lowerCamelCase_ : int =get_size_dict(snake_case__ , param_name="size" , default_to_square=snake_case__ ) lowerCamelCase_ : Optional[int] =resample if resample is not None else self.resample lowerCamelCase_ : List[str] =do_center_crop if do_center_crop is not None else self.do_center_crop lowerCamelCase_ : List[Any] =crop_size if crop_size is not None else self.crop_size lowerCamelCase_ : List[str] =get_size_dict(snake_case__ , param_name="crop_size" , default_to_square=snake_case__ ) lowerCamelCase_ : List[str] =do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase_ : List[str] =rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase_ : Optional[int] =do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase_ : Any =image_mean if image_mean is not None else self.image_mean lowerCamelCase_ : Tuple =image_std if image_std is not None else self.image_std lowerCamelCase_ : Dict =do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCamelCase_ : Dict =make_list_of_images(snake_case__ ) if not valid_images(snake_case__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCamelCase_ : Union[str, Any] =[convert_to_rgb(snake_case__ ) for image in images] # All transformations expect numpy arrays. lowerCamelCase_ : Union[str, Any] =[to_numpy_array(snake_case__ ) for image in images] if do_resize: lowerCamelCase_ : Dict =[self.resize(image=snake_case__ , size=snake_case__ , resample=snake_case__ ) for image in images] if do_center_crop: lowerCamelCase_ : Dict =[self.center_crop(image=snake_case__ , size=snake_case__ ) for image in images] if do_rescale: lowerCamelCase_ : List[str] =[self.rescale(image=snake_case__ , scale=snake_case__ ) for image in images] if do_normalize: lowerCamelCase_ : List[Any] =[self.normalize(image=snake_case__ , mean=snake_case__ , std=snake_case__ ) for image in images] lowerCamelCase_ : Dict =[to_channel_dimension_format(snake_case__ , snake_case__ ) for image in images] lowerCamelCase_ : int ={'''pixel_values''': images} return BatchFeature(data=snake_case__ , tensor_type=snake_case__ )
144
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __a (unittest.TestCase ): def __init__( self : int , __magic_name__ : Union[str, Any] , __magic_name__ : Tuple=3 , __magic_name__ : Tuple=32 , __magic_name__ : Optional[int]=3 , __magic_name__ : Optional[Any]=10 , __magic_name__ : str=[10, 20, 30, 40] , __magic_name__ : str=[1, 1, 2, 1] , __magic_name__ : Union[str, Any]=True , __magic_name__ : str=True , __magic_name__ : Dict="relu" , __magic_name__ : Any=3 , __magic_name__ : List[str]=None , ) -> List[Any]: """simple docstring""" UpperCAmelCase_ : str = parent UpperCAmelCase_ : List[str] = batch_size UpperCAmelCase_ : int = image_size UpperCAmelCase_ : Optional[int] = num_channels UpperCAmelCase_ : Any = embeddings_size UpperCAmelCase_ : Union[str, Any] = hidden_sizes UpperCAmelCase_ : Optional[Any] = depths UpperCAmelCase_ : Any = is_training UpperCAmelCase_ : Optional[Any] = use_labels UpperCAmelCase_ : List[Any] = hidden_act UpperCAmelCase_ : List[str] = num_labels UpperCAmelCase_ : Optional[Any] = scope UpperCAmelCase_ : int = len(__magic_name__ ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase_ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : Union[str, Any] = self.get_config() return config, pixel_values def UpperCAmelCase__ ( self : List[Any] ) -> Dict: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCAmelCase__ ( self : int , __magic_name__ : Union[str, Any] , __magic_name__ : str ) -> List[str]: """simple docstring""" UpperCAmelCase_ : Optional[Any] = FlaxRegNetModel(config=__magic_name__ ) UpperCAmelCase_ : Optional[Any] = model(__magic_name__ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase__ ( self : List[str] , __magic_name__ : str , __magic_name__ : Tuple ) -> List[Any]: """simple docstring""" UpperCAmelCase_ : Optional[int] = self.num_labels UpperCAmelCase_ : str = FlaxRegNetForImageClassification(config=__magic_name__ ) UpperCAmelCase_ : Tuple = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase_ : List[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ : str = config_and_inputs UpperCAmelCase_ : Optional[int] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class __a (lowerCamelCase , unittest.TestCase ): __a : Optional[Any] = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () __a : int = False __a : str = False __a : List[str] = False def UpperCAmelCase__ ( self : List[Any] ) -> None: """simple docstring""" UpperCAmelCase_ : Optional[int] = FlaxRegNetModelTester(self ) UpperCAmelCase_ : List[Any] = ConfigTester(self , config_class=__magic_name__ , has_text_modality=__magic_name__ ) def UpperCAmelCase__ ( self : List[Any] ) -> str: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: """simple docstring""" return def UpperCAmelCase__ ( self : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def UpperCAmelCase__ ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__magic_name__ ) @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def UpperCAmelCase__ ( self : Dict ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def UpperCAmelCase__ ( self : str ) -> List[str]: """simple docstring""" pass def UpperCAmelCase__ ( self : str ) -> Optional[int]: """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[int] = model_class(__magic_name__ ) UpperCAmelCase_ : List[Any] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : Dict = [*signature.parameters.keys()] UpperCAmelCase_ : Optional[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __magic_name__ ) def UpperCAmelCase__ ( self : List[str] ) -> List[str]: """simple docstring""" def check_hidden_states_output(__magic_name__ : Optional[int] , __magic_name__ : Union[str, Any] , __magic_name__ : Tuple ): UpperCAmelCase_ : Optional[int] = model_class(__magic_name__ ) UpperCAmelCase_ : str = model(**self._prepare_for_class(__magic_name__ , __magic_name__ ) ) UpperCAmelCase_ : Optional[int] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase_ : str = self.model_tester.num_stages self.assertEqual(len(__magic_name__ ) , expected_num_stages + 1 ) UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Any = True check_hidden_states_output(__magic_name__ , __magic_name__ , __magic_name__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : Any = True check_hidden_states_output(__magic_name__ , __magic_name__ , __magic_name__ ) def UpperCAmelCase__ ( self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase_ : str = self._prepare_for_class(__magic_name__ , __magic_name__ ) UpperCAmelCase_ : Optional[int] = model_class(__magic_name__ ) @jax.jit def model_jitted(__magic_name__ : Optional[int] , **__magic_name__ : Optional[Any] ): return model(pixel_values=__magic_name__ , **__magic_name__ ) with self.subTest('''JIT Enabled''' ): UpperCAmelCase_ : Dict = model_jitted(**__magic_name__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): UpperCAmelCase_ : Any = model_jitted(**__magic_name__ ).to_tuple() self.assertEqual(len(__magic_name__ ) , len(__magic_name__ ) ) for jitted_output, output in zip(__magic_name__ , __magic_name__ ): self.assertEqual(jitted_output.shape , output.shape ) def lowerCamelCase_ ( ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class __a (unittest.TestCase ): @cached_property def UpperCAmelCase__ ( self : List[str] ) -> Dict: """simple docstring""" return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''' ) if is_vision_available() else None @slow def UpperCAmelCase__ ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase_ : Dict = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''' ) UpperCAmelCase_ : Tuple = self.default_image_processor UpperCAmelCase_ : Tuple = prepare_img() UpperCAmelCase_ : Union[str, Any] = image_processor(images=__magic_name__ , return_tensors='''np''' ) UpperCAmelCase_ : int = model(**__magic_name__ ) # verify the logits UpperCAmelCase_ : int = (1, 10_00) self.assertEqual(outputs.logits.shape , __magic_name__ ) UpperCAmelCase_ : Optional[int] = jnp.array([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , __magic_name__ , atol=1E-4 ) )
125
0
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCAmelCase__ : Union[str, Any] ='pt' elif is_tf_available(): lowerCAmelCase__ : Dict ='tf' else: lowerCAmelCase__ : Dict ='jax' class __lowercase (__SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = ByTaTokenizer _UpperCAmelCase = False def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE_ : Optional[Any] = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__ ( self ): """simple docstring""" return ByTaTokenizer.from_pretrained('google/byt5-small' ) def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=2_0 , lowerCAmelCase__=5 ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = [] for i in range(len(lowerCAmelCase__ ) ): try: SCREAMING_SNAKE_CASE_ : str = tokenizer.decode([i] , clean_up_tokenization_spaces=lowerCAmelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) SCREAMING_SNAKE_CASE_ : List[str] = list(filter(lambda lowerCAmelCase__ : re.match(r'^[ a-zA-Z]+$' , t[1] ) , lowerCAmelCase__ ) ) SCREAMING_SNAKE_CASE_ : Any = list(filter(lambda lowerCAmelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=lowerCAmelCase__ ) , lowerCAmelCase__ ) ) if max_length is not None and len(lowerCAmelCase__ ) > max_length: SCREAMING_SNAKE_CASE_ : List[str] = toks[:max_length] if min_length is not None and len(lowerCAmelCase__ ) < min_length and len(lowerCAmelCase__ ) > 0: while len(lowerCAmelCase__ ) < min_length: SCREAMING_SNAKE_CASE_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE_ : Any = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.decode(lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ ) if " " not in output_txt and len(lowerCAmelCase__ ) > 1: SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowerCAmelCase__ ) + ' ' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowerCAmelCase__ ) ) if with_prefix_space: SCREAMING_SNAKE_CASE_ : List[Any] = ' ' + output_txt SCREAMING_SNAKE_CASE_ : str = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) return output_txt, output_ids def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : str = tokenizer(['hi</s>', 'I went to the gym</s>', '</s>'] ) SCREAMING_SNAKE_CASE_ : List[str] = tokenizer(['hi', 'I went to the gym', ''] ) self.assertListEqual(batch_with_eos_added['input_ids'] , batch_without_eos_added['input_ids'] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Optional[int] = 'Unicode €.' SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = [8_8, 1_1_3, 1_0_8, 1_0_2, 1_1_4, 1_0_3, 1_0_4, 3_5, 2_2_9, 1_3_3, 1_7_5, 4_9, 1] self.assertEqual(encoded['input_ids'] , lowerCAmelCase__ ) # decoding SCREAMING_SNAKE_CASE_ : str = tokenizer.decode(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , 'Unicode €.</s>' ) SCREAMING_SNAKE_CASE_ : Any = tokenizer('e è é ê ë' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = [1_0_4, 3_5, 1_9_8, 1_7_1, 3_5, 1_9_8, 1_7_2, 3_5, 1_9_8, 1_7_3, 3_5, 1_9_8, 1_7_4, 1] self.assertEqual(encoded['input_ids'] , lowerCAmelCase__ ) # decoding SCREAMING_SNAKE_CASE_ : str = tokenizer.decode(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , 'e è é ê ë</s>' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ) , 'e è é ê ë</s>' ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Optional[int] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off SCREAMING_SNAKE_CASE_ : int = [6_8, 3_5, 1_1_1, 1_1_4, 1_1_3, 1_0_6, 3_5, 1_1_5, 1_0_0, 1_1_7, 1_0_0, 1_0_6, 1_1_7, 1_0_0, 1_1_5, 1_0_7, 3_5, 1_0_5, 1_1_4, 1_1_7, 3_5, 1_1_8, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_0_8, 1_2_5, 1_0_0, 1_1_9, 1_0_8, 1_1_4, 1_1_3, 4_9, 1, 0] # fmt: on SCREAMING_SNAKE_CASE_ : str = tokenizer(lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) if FRAMEWORK != "jax": SCREAMING_SNAKE_CASE_ : Any = list(batch.input_ids.numpy()[0] ) else: SCREAMING_SNAKE_CASE_ : List[Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual((2, 3_7) , batch.input_ids.shape ) self.assertEqual((2, 3_7) , batch.attention_mask.shape ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : int = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] SCREAMING_SNAKE_CASE_ : Dict = tokenizer(lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids' , lowerCAmelCase__ ) self.assertIn('attention_mask' , lowerCAmelCase__ ) self.assertNotIn('decoder_input_ids' , lowerCAmelCase__ ) self.assertNotIn('decoder_attention_mask' , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Union[str, Any] = [ 'Summary of the text.', 'Another summary.', ] SCREAMING_SNAKE_CASE_ : List[str] = tokenizer( text_target=lowerCAmelCase__ , max_length=3_2 , padding='max_length' , truncation=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) self.assertEqual(3_2 , targets['input_ids'].shape[1] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Tuple = ['A long paragraph for summarization. </s>'] SCREAMING_SNAKE_CASE_ : Any = ['Summary of the text. </s>'] # fmt: off SCREAMING_SNAKE_CASE_ : Dict = [6_8, 3_5, 1_1_1, 1_1_4, 1_1_3, 1_0_6, 3_5, 1_1_5, 1_0_0, 1_1_7, 1_0_0, 1_0_6, 1_1_7, 1_0_0, 1_1_5, 1_0_7, 3_5, 1_0_5, 1_1_4, 1_1_7, 3_5, 1_1_8, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_0_8, 1_2_5, 1_0_0, 1_1_9, 1_0_8, 1_1_4, 1_1_3, 4_9, 3_5, 1] SCREAMING_SNAKE_CASE_ : List[Any] = [8_6, 1_2_0, 1_1_2, 1_1_2, 1_0_0, 1_1_7, 1_2_4, 3_5, 1_1_4, 1_0_5, 3_5, 1_1_9, 1_0_7, 1_0_4, 3_5, 1_1_9, 1_0_4, 1_2_3, 1_1_9, 4_9, 3_5, 1] # fmt: on SCREAMING_SNAKE_CASE_ : Tuple = tokenizer(lowerCAmelCase__ , text_target=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , batch['input_ids'][0] ) self.assertEqual(lowerCAmelCase__ , batch['labels'][0] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ : Union[str, Any] = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : Any = ' He is very happy, UNwant\u00E9d,running' SCREAMING_SNAKE_CASE_ : Any = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) tokenizer.save_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer.__class__.from_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = after_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) shutil.rmtree(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ : Dict = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : Optional[int] = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) SCREAMING_SNAKE_CASE_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) tokenizer.save_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = tokenizer.__class__.from_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = after_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertIn('new_additional_special_token' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.__class__.from_pretrained(lowerCAmelCase__ , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , 'special_tokens_map.json' ) , encoding='utf-8' ) as json_file: SCREAMING_SNAKE_CASE_ : int = json.load(lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , 'tokenizer_config.json' ) , encoding='utf-8' ) as json_file: SCREAMING_SNAKE_CASE_ : Optional[int] = json.load(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[str] = [F'''<extra_id_{i}>''' for i in range(1_2_5 )] SCREAMING_SNAKE_CASE_ : List[Any] = added_tokens_extra_ids + [ 'an_additional_special_token' ] SCREAMING_SNAKE_CASE_ : List[str] = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(lowerCAmelCase__ , 'special_tokens_map.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , 'tokenizer_config.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE_ : Any = tokenizer_class.from_pretrained( lowerCAmelCase__ , ) self.assertIn( 'an_additional_special_token' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['an_additional_special_token'] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE_ : int = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token' , lstrip=lowerCAmelCase__ )] SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer_class.from_pretrained( lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , ) self.assertIn('a_new_additional_special_token' , tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ) , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer_class.from_pretrained(lowerCAmelCase__ ) self.assertTrue(tokenizer.decode([2_5_5] ) == '' ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.get_tokenizers(fast=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ : Tuple = ['t', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 'x', 't', '</s>'] SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ : Tuple = [ 'bos_token', 'eos_token', 'unk_token', 'sep_token', 'pad_token', 'cls_token', 'mask_token', ] SCREAMING_SNAKE_CASE_ : Union[str, Any] = 0 SCREAMING_SNAKE_CASE_ : Tuple = tokenizer.convert_ids_to_tokens( lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) for attr in attributes_list: setattr(lowerCAmelCase__ , attr + '_id' , lowerCAmelCase__ ) self.assertEqual(getattr(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(getattr(lowerCAmelCase__ , attr + '_id' ) , lowerCAmelCase__ ) setattr(lowerCAmelCase__ , attr + '_id' , lowerCAmelCase__ ) self.assertEqual(getattr(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(getattr(lowerCAmelCase__ , attr + '_id' ) , lowerCAmelCase__ ) setattr(lowerCAmelCase__ , 'additional_special_tokens_ids' , [] ) self.assertListEqual(getattr(lowerCAmelCase__ , 'additional_special_tokens' ) , [] ) self.assertListEqual(getattr(lowerCAmelCase__ , 'additional_special_tokens_ids' ) , [] ) setattr(lowerCAmelCase__ , 'additional_special_tokens_ids' , [token_id_to_test_setters] ) self.assertListEqual(getattr(lowerCAmelCase__ , 'additional_special_tokens' ) , [token_to_test_setters] ) self.assertListEqual(getattr(lowerCAmelCase__ , 'additional_special_tokens_ids' ) , [token_id_to_test_setters] )
162
import socket def a__ ( ): SCREAMING_SNAKE_CASE_ : Dict = socket.socket(socket.AF_INET, socket.SOCK_STREAM ) SCREAMING_SNAKE_CASE_ : Any = socket.gethostname() SCREAMING_SNAKE_CASE_ : List[str] = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file', 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: SCREAMING_SNAKE_CASE_ : Tuple = sock.recv(1_0_2_4 ) if not data: break out_file.write(A__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
162
1
from ..utils import DummyObject, requires_backends class _snake_case ( metaclass=__a ): '''simple docstring''' A__ : Any = ["note_seq"] def __init__( self: Dict ,*lowerCamelCase_: Union[str, Any] ,**lowerCamelCase_: List[str] ) -> List[Any]: requires_backends(self ,["""note_seq"""] ) @classmethod def A__ ( cls: Optional[int] ,*lowerCamelCase_: Any ,**lowerCamelCase_: Optional[Any] ) -> str: requires_backends(cls ,["""note_seq"""] ) @classmethod def A__ ( cls: Tuple ,*lowerCamelCase_: Optional[Any] ,**lowerCamelCase_: List[str] ) -> Optional[int]: requires_backends(cls ,["""note_seq"""] )
345
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
0
import random def SCREAMING_SNAKE_CASE ( snake_case_ : int , snake_case_ : float , snake_case_ : bool = False ): snake_case__ : dict = {i: [] for i in range(snake_case_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(snake_case_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(snake_case_ ): for j in range(i + 1 , snake_case_ ): if random.random() < probability: graph[i].append(snake_case_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(snake_case_ ) return graph def SCREAMING_SNAKE_CASE ( snake_case_ : int ): return { i: [j for j in range(snake_case_ ) if i != j] for i in range(snake_case_ ) } if __name__ == "__main__": import doctest doctest.testmod()
286
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width", "cross_attention_kwargs"} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def _lowercase ( self : List[str] ): torch.manual_seed(0 ) snake_case__ : Any = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=3_2 , ) snake_case__ : int = PNDMScheduler(skip_prk_steps=__A ) torch.manual_seed(0 ) snake_case__ : Union[str, Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case__ : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) snake_case__ : Union[str, Any] = CLIPTextModel(__A ) snake_case__ : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) snake_case__ : str = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def _lowercase ( self : List[Any] , __A : int , __A : Any=0 ): snake_case__ : Optional[int] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__A ) ).to(__A ) snake_case__ : int = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case__ : Union[str, Any] = Image.fromarray(np.uinta(__A ) ).convert("RGB" ) if str(__A ).startswith("mps" ): snake_case__ : List[Any] = torch.manual_seed(__A ) else: snake_case__ : Optional[int] = torch.Generator(device=__A ).manual_seed(__A ) snake_case__ : Optional[int] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "image_guidance_scale": 1, "output_type": "numpy", } return inputs def _lowercase ( self : int ): snake_case__ : Tuple = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case__ : int = self.get_dummy_components() snake_case__ : List[Any] = StableDiffusionInstructPixaPixPipeline(**__A ) snake_case__ : List[Any] = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) snake_case__ : Tuple = self.get_dummy_inputs(__A ) snake_case__ : List[str] = sd_pipe(**__A ).images snake_case__ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : List[Any] = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self : Union[str, Any] ): snake_case__ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case__ : List[Any] = self.get_dummy_components() snake_case__ : List[Any] = StableDiffusionInstructPixaPixPipeline(**__A ) snake_case__ : str = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) snake_case__ : str = self.get_dummy_inputs(__A ) snake_case__ : List[Any] = "french fries" snake_case__ : str = sd_pipe(**__A , negative_prompt=__A ) snake_case__ : Any = output.images snake_case__ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : Union[str, Any] = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self : Optional[int] ): snake_case__ : Tuple = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case__ : List[Any] = self.get_dummy_components() snake_case__ : str = StableDiffusionInstructPixaPixPipeline(**__A ) snake_case__ : List[str] = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) snake_case__ : Any = self.get_dummy_inputs(__A ) snake_case__ : Tuple = [inputs["prompt"]] * 2 snake_case__ : Any = np.array(inputs["image"] ).astype(np.floataa ) / 2_5_5.0 snake_case__ : List[str] = torch.from_numpy(__A ).unsqueeze(0 ).to(__A ) snake_case__ : Union[str, Any] = image / 2 + 0.5 snake_case__ : str = image.permute(0 , 3 , 1 , 2 ) snake_case__ : int = image.repeat(2 , 1 , 1 , 1 ) snake_case__ : str = sd_pipe(**__A ).images snake_case__ : Any = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) snake_case__ : int = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self : Union[str, Any] ): snake_case__ : Union[str, Any] = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case__ : int = self.get_dummy_components() snake_case__ : Dict = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" ) snake_case__ : Tuple = StableDiffusionInstructPixaPixPipeline(**__A ) snake_case__ : str = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) snake_case__ : str = self.get_dummy_inputs(__A ) snake_case__ : Optional[Any] = sd_pipe(**__A ).images snake_case__ : Dict = image[0, -3:, -3:, -1] snake_case__ : Union[str, Any] = [round(__A , 4 ) for x in image_slice.flatten().tolist()] print(",".join([str(__A ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) snake_case__ : str = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self : List[str] ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def _lowercase ( self : List[Any] ): snake_case__ : Tuple = self.get_dummy_components() snake_case__ : Tuple = StableDiffusionInstructPixaPixPipeline(**__A ) snake_case__ : int = VaeImageProcessor(do_resize=__A , do_normalize=__A ) snake_case__ : Any = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) snake_case__ : Dict = pipe(**self.get_dummy_inputs_by_type(__A , input_image_type="pt" ) )[0] snake_case__ : int = components["vae"] snake_case__ : Union[str, Any] = self.get_dummy_inputs_by_type(__A , input_image_type="pt" ) for image_param in self.image_latents_params: if image_param in inputs.keys(): snake_case__ : Optional[int] = vae.encode(inputs[image_param] ).latent_dist.mode() snake_case__ : str = pipe(**__A )[0] snake_case__ : Dict = np.abs(out - out_latents_inputs ).max() self.assertLess(__A , 1e-4 , "passing latents as image input generate different result from passing image" ) @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Optional[int] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : str , __A : Dict=0 ): snake_case__ : Optional[int] = torch.manual_seed(__A ) snake_case__ : Tuple = load_image( "https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg" ) snake_case__ : Optional[Any] = { "prompt": "turn him into a cyborg", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "image_guidance_scale": 1.0, "output_type": "numpy", } return inputs def _lowercase ( self : int ): snake_case__ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() snake_case__ : Union[str, Any] = self.get_inputs() snake_case__ : Union[str, Any] = pipe(**__A ).images snake_case__ : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : Any = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def _lowercase ( self : str ): snake_case__ : int = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=__A ) snake_case__ : Dict = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() snake_case__ : List[str] = self.get_inputs() snake_case__ : Any = pipe(**__A ).images snake_case__ : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : Optional[Any] = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def _lowercase ( self : Dict ): snake_case__ : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=__A ) snake_case__ : List[str] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() snake_case__ : int = self.get_inputs() snake_case__ : Union[str, Any] = pipe(**__A ).images snake_case__ : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) snake_case__ : Union[str, Any] = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def _lowercase ( self : List[Any] ): snake_case__ : Optional[Any] = 0 def callback_fn(__A : int , __A : int , __A : torch.FloatTensor ) -> None: snake_case__ : Union[str, Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: snake_case__ : Optional[Any] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) snake_case__ : int = latents[0, -3:, -3:, -1] snake_case__ : Optional[int] = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: snake_case__ : int = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) snake_case__ : Any = latents[0, -3:, -3:, -1] snake_case__ : Dict = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 snake_case__ : Any = False snake_case__ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=__A , torch_dtype=torch.floataa ) snake_case__ : int = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() snake_case__ : Optional[Any] = self.get_inputs() pipe(**__A , callback=__A , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def _lowercase ( self : List[Any] ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case__ : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=__A , torch_dtype=torch.floataa ) snake_case__ : Tuple = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case__ : Dict = self.get_inputs() snake_case__ : List[Any] = pipe(**__A ) snake_case__ : List[str] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def _lowercase ( self : Tuple ): snake_case__ : int = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 snake_case__ : Union[str, Any] = inputs["image"].resize((5_0_4, 5_0_4) ) snake_case__ : Optional[Any] = "timbrooks/instruct-pix2pix" snake_case__ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( __A , safety_checker=__A , ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() snake_case__ : Union[str, Any] = pipe(**__A ) snake_case__ : Tuple = output.images[0] snake_case__ : List[Any] = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) snake_case__ : int = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
286
1
import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib lowerCAmelCase = { '''debug''': logging.DEBUG, '''info''': logging.INFO, '''warning''': logging.WARNING, '''error''': logging.ERROR, '''critical''': logging.CRITICAL, } lowerCAmelCase = logging.WARNING def _lowerCamelCase( ) -> Dict: '''simple docstring''' __lowercase= os.getenv('DATASETS_VERBOSITY' , lowercase__ ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F'Unknown option DATASETS_VERBOSITY={env_level_str}, ' F'has to be one of: { ", ".join(log_levels.keys() ) }' ) return _default_log_level def _lowerCamelCase( ) -> str: '''simple docstring''' return __name__.split('.' )[0] def _lowerCamelCase( ) -> logging.Logger: '''simple docstring''' return logging.getLogger(_get_library_name() ) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level() ) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET ) def _lowerCamelCase( lowercase__ = None ) -> logging.Logger: '''simple docstring''' if name is None: __lowercase= _get_library_name() return logging.getLogger(lowercase__ ) def _lowerCamelCase( ) -> int: '''simple docstring''' return _get_library_root_logger().getEffectiveLevel() def _lowerCamelCase( lowercase__ ) -> None: '''simple docstring''' _get_library_root_logger().setLevel(lowercase__ ) def _lowerCamelCase( ) -> Any: '''simple docstring''' return set_verbosity(lowercase__ ) def _lowerCamelCase( ) -> Optional[int]: '''simple docstring''' return set_verbosity(lowercase__ ) def _lowerCamelCase( ) -> Optional[Any]: '''simple docstring''' return set_verbosity(lowercase__ ) def _lowerCamelCase( ) -> str: '''simple docstring''' return set_verbosity(lowercase__ ) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= False def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class A : def __init__(self , *lowerCAmelCase , **lowerCAmelCase ): # pylint: disable=unused-argument __lowercase= args[0] if args else None def __iter__(self ): return iter(self._iterator ) def __getattr__(self , lowerCAmelCase ): def empty_fn(*lowerCAmelCase , **lowerCAmelCase ): # pylint: disable=unused-argument return return empty_fn def __enter__(self ): return self def __exit__(self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): return lowerCAmelCase = True class A : def __call__(self , *lowerCAmelCase , lowerCAmelCase=False , **lowerCAmelCase ): if _tqdm_active and not disable: return tqdm_lib.tqdm(*lowerCAmelCase , **lowerCAmelCase ) else: return EmptyTqdm(*lowerCAmelCase , **lowerCAmelCase ) def _A (self , *lowerCAmelCase , **lowerCAmelCase ): __lowercase= None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*lowerCAmelCase , **lowerCAmelCase ) def _A (self ): if _tqdm_active: return tqdm_lib.tqdm.get_lock() lowerCAmelCase = _tqdm_cls() def _lowerCamelCase( ) -> bool: '''simple docstring''' global _tqdm_active return bool(_tqdm_active ) def _lowerCamelCase( ) -> Any: '''simple docstring''' global _tqdm_active __lowercase= True def _lowerCamelCase( ) -> Dict: '''simple docstring''' global _tqdm_active __lowercase= False
295
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class A ( A_ ): def __init__(self , lowerCAmelCase , lowerCAmelCase=1_3 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=True , lowerCAmelCase=9_9 , lowerCAmelCase=3_2 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=3_7 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=5_1_2 , lowerCAmelCase=1_6 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=None , ): __lowercase= parent __lowercase= batch_size __lowercase= seq_length __lowercase= is_training __lowercase= use_input_mask __lowercase= use_token_type_ids __lowercase= use_labels __lowercase= vocab_size __lowercase= hidden_size __lowercase= num_hidden_layers __lowercase= num_attention_heads __lowercase= intermediate_size __lowercase= hidden_act __lowercase= hidden_dropout_prob __lowercase= attention_probs_dropout_prob __lowercase= max_position_embeddings __lowercase= type_vocab_size __lowercase= type_sequence_label_size __lowercase= initializer_range __lowercase= num_labels __lowercase= num_choices __lowercase= scope def _A (self ): __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase= None if self.use_input_mask: __lowercase= random_attention_mask([self.batch_size, self.seq_length] ) __lowercase= None __lowercase= None __lowercase= None if self.use_labels: __lowercase= ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase= ids_tensor([self.batch_size] , self.num_choices ) __lowercase= self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _A (self ): return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= DistilBertModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= model(lowerCAmelCase , lowerCAmelCase ) __lowercase= model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= DistilBertForMaskedLM(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= DistilBertForQuestionAnswering(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= model( lowerCAmelCase , attention_mask=lowerCAmelCase , start_positions=lowerCAmelCase , end_positions=lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= self.num_labels __lowercase= DistilBertForSequenceClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= self.num_labels __lowercase= DistilBertForTokenClassification(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= self.num_choices __lowercase= DistilBertForMultipleChoice(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() __lowercase= input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase= input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase= model( lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _A (self ): __lowercase= self.prepare_config_and_inputs() ((__lowercase), (__lowercase), (__lowercase), (__lowercase), (__lowercase), (__lowercase))= config_and_inputs __lowercase= {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A ( A_ , A_ , unittest.TestCase ): UpperCamelCase_ : Any =( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) UpperCamelCase_ : Optional[int] =( { '''feature-extraction''': DistilBertModel, '''fill-mask''': DistilBertForMaskedLM, '''question-answering''': DistilBertForQuestionAnswering, '''text-classification''': DistilBertForSequenceClassification, '''token-classification''': DistilBertForTokenClassification, '''zero-shot''': DistilBertForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase_ : str =True UpperCamelCase_ : str =True UpperCamelCase_ : Union[str, Any] =True UpperCamelCase_ : Optional[int] =True def _A (self ): __lowercase= DistilBertModelTester(self ) __lowercase= ConfigTester(self , config_class=lowerCAmelCase , dim=3_7 ) def _A (self ): self.config_tester.run_common_tests() def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*lowerCAmelCase ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*lowerCAmelCase ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*lowerCAmelCase ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*lowerCAmelCase ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*lowerCAmelCase ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*lowerCAmelCase ) @slow def _A (self ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase= DistilBertModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @slow @require_torch_gpu def _A (self ): __lowercase, __lowercase= self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase= True __lowercase= model_class(config=lowerCAmelCase ) __lowercase= self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) __lowercase= torch.jit.trace( lowerCAmelCase , (inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCAmelCase , os.path.join(lowerCAmelCase , 'traced_model.pt' ) ) __lowercase= torch.jit.load(os.path.join(lowerCAmelCase , 'traced_model.pt' ) , map_location=lowerCAmelCase ) loaded(inputs_dict['input_ids'].to(lowerCAmelCase ) , inputs_dict['attention_mask'].to(lowerCAmelCase ) ) @require_torch class A ( unittest.TestCase ): @slow def _A (self ): __lowercase= DistilBertModel.from_pretrained('distilbert-base-uncased' ) __lowercase= torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowercase= torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase= model(lowerCAmelCase , attention_mask=lowerCAmelCase )[0] __lowercase= torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , lowerCAmelCase ) __lowercase= torch.tensor( [[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase , atol=1E-4 ) )
295
1
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCamelCase_ = logging.get_logger(__name__) def A ( __UpperCAmelCase , __UpperCAmelCase ) -> Any: '''simple docstring''' try: with open(__UpperCAmelCase , '''rb''' ) as flax_state_f: UpperCAmelCase_ = from_bytes(__UpperCAmelCase , flax_state_f.read() ) except UnpicklingError as e: try: with open(__UpperCAmelCase ) as f: if f.read().startswith('''version''' ): raise OSError( '''You seem to have cloned a repository without having git-lfs installed. Please''' ''' install git-lfs and run `git lfs install` followed by `git lfs pull` in the''' ''' folder you cloned.''' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(__UpperCAmelCase , __UpperCAmelCase ) def A ( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' try: import torch # noqa: F401 except ImportError: logger.error( '''Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights UpperCAmelCase_ = flatten_dict(jax.tree_util.tree_map(lambda __UpperCAmelCase : x.dtype == jnp.bfloataa , __UpperCAmelCase ) ).values() if any(__UpperCAmelCase ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) UpperCAmelCase_ = jax.tree_util.tree_map( lambda __UpperCAmelCase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , __UpperCAmelCase ) UpperCAmelCase_ = '''''' UpperCAmelCase_ = flatten_dict(__UpperCAmelCase , sep='''.''' ) UpperCAmelCase_ = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase_ = [] UpperCAmelCase_ = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase_ = flax_key_tuple.split('''.''' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase_ = flax_key_tuple_array[:-1] + ['''weight'''] UpperCAmelCase_ = jnp.transpose(__UpperCAmelCase , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase_ = flax_key_tuple_array[:-1] + ['''weight'''] UpperCAmelCase_ = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase_ = flax_key_tuple_array[:-1] + ['''weight'''] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(__UpperCAmelCase ): UpperCAmelCase_ = ( flax_key_tuple_string.replace('''_0''' , '''.0''' ) .replace('''_1''' , '''.1''' ) .replace('''_2''' , '''.2''' ) .replace('''_3''' , '''.3''' ) .replace('''_4''' , '''.4''' ) .replace('''_5''' , '''.5''' ) .replace('''_6''' , '''.6''' ) .replace('''_7''' , '''.7''' ) .replace('''_8''' , '''.8''' ) .replace('''_9''' , '''.9''' ) ) UpperCAmelCase_ = '''.'''.join(__UpperCAmelCase ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " f"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict UpperCAmelCase_ = np.asarray(__UpperCAmelCase ) if not isinstance(__UpperCAmelCase , np.ndarray ) else flax_tensor UpperCAmelCase_ = torch.from_numpy(__UpperCAmelCase ) # remove from missing keys missing_keys.remove(__UpperCAmelCase ) else: # weight is not expected by PyTorch model unexpected_keys.append(__UpperCAmelCase ) pt_model.load_state_dict(__UpperCAmelCase ) # re-transform missing_keys to list UpperCAmelCase_ = list(__UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' f" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" f" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' f" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) if len(__UpperCAmelCase ) > 0: logger.warning( f"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" f" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" ''' use it for predictions and inference.''' ) return pt_model
344
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "s-JoL/Open-Llama-V1": "https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json", } class a_ ( _snake_case ): UpperCamelCase__ : Dict ="open-llama" def __init__( self :Union[str, Any] , _lowercase :List[Any]=100000 , _lowercase :Dict=4096 , _lowercase :List[Any]=11008 , _lowercase :Optional[int]=32 , _lowercase :Union[str, Any]=32 , _lowercase :List[str]="silu" , _lowercase :Union[str, Any]=2048 , _lowercase :Any=0.02 , _lowercase :Optional[Any]=1E-6 , _lowercase :str=True , _lowercase :str=0 , _lowercase :Any=1 , _lowercase :Optional[Any]=2 , _lowercase :str=False , _lowercase :Dict=True , _lowercase :Optional[Any]=0.1 , _lowercase :Tuple=0.1 , _lowercase :Dict=True , _lowercase :List[Any]=True , _lowercase :Dict=None , **_lowercase :Optional[int] , ) -> List[Any]: UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = initializer_range UpperCAmelCase_ = rms_norm_eps UpperCAmelCase_ = use_cache UpperCAmelCase_ = kwargs.pop( '''use_memorry_efficient_attention''' , _lowercase) UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_dropout_prob UpperCAmelCase_ = use_stable_embedding UpperCAmelCase_ = shared_input_output_embedding UpperCAmelCase_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , tie_word_embeddings=_lowercase , **_lowercase , ) def __a ( self :int) -> str: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _lowercase) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f"got {self.rope_scaling}") UpperCAmelCase_ = self.rope_scaling.get('''type''' , _lowercase) UpperCAmelCase_ = self.rope_scaling.get('''factor''' , _lowercase) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}") if rope_scaling_factor is None or not isinstance(_lowercase , _lowercase) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}")
344
1
'''simple docstring''' import os def __a(SCREAMING_SNAKE_CASE_ : Any ): '''simple docstring''' _lowerCAmelCase = len(grid[0] ) _lowerCAmelCase = len(_UpperCAmelCase ) _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(_UpperCAmelCase ): for j in range(n_rows - 3 ): _lowerCAmelCase = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] _lowerCAmelCase = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: _lowerCAmelCase = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: _lowerCAmelCase = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) _lowerCAmelCase = max( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if max_product > largest: _lowerCAmelCase = max_product return largest def __a(): '''simple docstring''' _lowerCAmelCase = [] with open(os.path.dirname(_UpperCAmelCase ) + "/grid.txt" ) as file: for line in file: grid.append(line.strip("\n" ).split(" " ) ) _lowerCAmelCase = [[int(_UpperCAmelCase ) for i in grid[j]] for j in range(len(_UpperCAmelCase ) )] return largest_product(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
158
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase : Tuple = { '''configuration_lilt''': ['''LILT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LiltConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : List[Any] = [ '''LILT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LiltForQuestionAnswering''', '''LiltForSequenceClassification''', '''LiltForTokenClassification''', '''LiltModel''', '''LiltPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys __UpperCamelCase : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
309
"""simple docstring""" from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __UpperCamelCase : str = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __UpperCamelCase : Optional[Any] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __UpperCamelCase : Dict = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : str ): lowerCAmelCase = len([g for position, g in enumerate(_UpperCAmelCase ) if g == main_target[position]] ) return (item, float(_UpperCAmelCase )) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : str ): lowerCAmelCase = random.randint(0 , len(_UpperCAmelCase ) - 1 ) lowerCAmelCase = parent_a[:random_slice] + parent_a[random_slice:] lowerCAmelCase = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] ): lowerCAmelCase = list(_UpperCAmelCase ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: lowerCAmelCase = random.choice(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : tuple[str, float] , _UpperCAmelCase : list[tuple[str, float]] , _UpperCAmelCase : list[str] , ): lowerCAmelCase = [] # Generate more children proportionally to the fitness score. lowerCAmelCase = int(parent_a[1] * 100 ) + 1 lowerCAmelCase = 10 if child_n >= 10 else child_n for _ in range(_UpperCAmelCase ): lowerCAmelCase = population_score[random.randint(0 , _UpperCAmelCase )][0] lowerCAmelCase ,lowerCAmelCase = crossover(parent_a[0] , _UpperCAmelCase ) # Append new string to the population list. pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) return pop def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] , _UpperCAmelCase : bool = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: lowerCAmelCase = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(_UpperCAmelCase ) # Verify that the target contains no genes besides the ones inside genes variable. lowerCAmelCase = sorted({c for c in target if c not in genes} ) if not_in_genes_list: lowerCAmelCase = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(_UpperCAmelCase ) # Generate random starting population. lowerCAmelCase = [] for _ in range(_UpperCAmelCase ): population.append(''.join([random.choice(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) )] ) ) # Just some logs to know what the algorithms is doing. lowerCAmelCase ,lowerCAmelCase = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_UpperCAmelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. lowerCAmelCase = [evaluate(_UpperCAmelCase , _UpperCAmelCase ) for item in population] # Check if there is a matching evolution. lowerCAmelCase = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x[1] , reverse=_UpperCAmelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. lowerCAmelCase = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_UpperCAmelCase ) # Normalize population score to be between 0 and 1. lowerCAmelCase = [ (item, score / len(_UpperCAmelCase )) for item, score in population_score ] # This is selection for i in range(_UpperCAmelCase ): population.extend(select(population_score[int(_UpperCAmelCase )] , _UpperCAmelCase , _UpperCAmelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_UpperCAmelCase ) > N_POPULATION: break if __name__ == "__main__": __UpperCamelCase : Tuple = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __UpperCamelCase : str = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase : Dict = basic(target_str, genes_list) print( f'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
309
1
'''simple docstring''' from statistics import mean, stdev def A__ ( UpperCAmelCase_ , UpperCAmelCase_ = 3 ): _UpperCamelCase : List[Any] = min(UpperCAmelCase_ ) _UpperCamelCase : Dict = max(UpperCAmelCase_ ) # normalize data return [round((x - x_min) / (x_max - x_min) , UpperCAmelCase_ ) for x in data] def A__ ( UpperCAmelCase_ , UpperCAmelCase_ = 3 ): _UpperCamelCase : str = mean(UpperCAmelCase_ ) _UpperCamelCase : Tuple = stdev(UpperCAmelCase_ ) # standardize data return [round((x - mu) / (sigma) , UpperCAmelCase_ ) for x in data]
83
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
1
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase ( unittest.TestCase ): def a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'bird' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) snake_case_ = pipe.prepare_image_inputs([canny_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'Chef in the kitchen' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) snake_case_ = pipe.prepare_image_inputs([pose_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
371
from __future__ import annotations import os from typing import Any import requests _UpperCAmelCase : int = """https://api.github.com""" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _UpperCAmelCase : Dict = BASE_URL + """/user""" # https://github.com/settings/tokens _UpperCAmelCase : Optional[Any] = os.environ.get("""USER_TOKEN""", """""") def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' snake_case_ = { 'Authorization': F'''token {auth_token}''', 'Accept': 'application/vnd.github.v3+json', } return requests.get(UpperCamelCase__ , headers=UpperCamelCase__ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F'''{key}: {value}''') else: raise ValueError("""'USER_TOKEN' field cannot be empty.""")
200
0
'''simple docstring''' import argparse import struct import unittest class _lowerCamelCase : '''simple docstring''' def __init__( self : int , _A : bytes ) -> None: __magic_name__ : Dict = data # Initialize hash values __magic_name__ : Optional[int] = [ 0X6_A_0_9_E_6_6_7, 0XB_B_6_7_A_E_8_5, 0X3_C_6_E_F_3_7_2, 0XA_5_4_F_F_5_3_A, 0X5_1_0_E_5_2_7_F, 0X9_B_0_5_6_8_8_C, 0X1_F_8_3_D_9_A_B, 0X5_B_E_0_C_D_1_9, ] # Initialize round constants __magic_name__ : Dict = [ 0X4_2_8_A_2_F_9_8, 0X7_1_3_7_4_4_9_1, 0XB_5_C_0_F_B_C_F, 0XE_9_B_5_D_B_A_5, 0X3_9_5_6_C_2_5_B, 0X5_9_F_1_1_1_F_1, 0X9_2_3_F_8_2_A_4, 0XA_B_1_C_5_E_D_5, 0XD_8_0_7_A_A_9_8, 0X1_2_8_3_5_B_0_1, 0X2_4_3_1_8_5_B_E, 0X5_5_0_C_7_D_C_3, 0X7_2_B_E_5_D_7_4, 0X8_0_D_E_B_1_F_E, 0X9_B_D_C_0_6_A_7, 0XC_1_9_B_F_1_7_4, 0XE_4_9_B_6_9_C_1, 0XE_F_B_E_4_7_8_6, 0X0_F_C_1_9_D_C_6, 0X2_4_0_C_A_1_C_C, 0X2_D_E_9_2_C_6_F, 0X4_A_7_4_8_4_A_A, 0X5_C_B_0_A_9_D_C, 0X7_6_F_9_8_8_D_A, 0X9_8_3_E_5_1_5_2, 0XA_8_3_1_C_6_6_D, 0XB_0_0_3_2_7_C_8, 0XB_F_5_9_7_F_C_7, 0XC_6_E_0_0_B_F_3, 0XD_5_A_7_9_1_4_7, 0X0_6_C_A_6_3_5_1, 0X1_4_2_9_2_9_6_7, 0X2_7_B_7_0_A_8_5, 0X2_E_1_B_2_1_3_8, 0X4_D_2_C_6_D_F_C, 0X5_3_3_8_0_D_1_3, 0X6_5_0_A_7_3_5_4, 0X7_6_6_A_0_A_B_B, 0X8_1_C_2_C_9_2_E, 0X9_2_7_2_2_C_8_5, 0XA_2_B_F_E_8_A_1, 0XA_8_1_A_6_6_4_B, 0XC_2_4_B_8_B_7_0, 0XC_7_6_C_5_1_A_3, 0XD_1_9_2_E_8_1_9, 0XD_6_9_9_0_6_2_4, 0XF_4_0_E_3_5_8_5, 0X1_0_6_A_A_0_7_0, 0X1_9_A_4_C_1_1_6, 0X1_E_3_7_6_C_0_8, 0X2_7_4_8_7_7_4_C, 0X3_4_B_0_B_C_B_5, 0X3_9_1_C_0_C_B_3, 0X4_E_D_8_A_A_4_A, 0X5_B_9_C_C_A_4_F, 0X6_8_2_E_6_F_F_3, 0X7_4_8_F_8_2_E_E, 0X7_8_A_5_6_3_6_F, 0X8_4_C_8_7_8_1_4, 0X8_C_C_7_0_2_0_8, 0X9_0_B_E_F_F_F_A, 0XA_4_5_0_6_C_E_B, 0XB_E_F_9_A_3_F_7, 0XC_6_7_1_7_8_F_2, ] __magic_name__ : str = self.preprocessing(self.data ) self.final_hash() @staticmethod def __lowerCAmelCase ( _A : bytes ) -> bytes: __magic_name__ : int = b'\x80' + (b'\x00' * (63 - (len(_A ) + 8) % 64)) __magic_name__ : Optional[int] = struct.pack('>Q' , (len(_A ) * 8) ) return data + padding + big_endian_integer def __lowerCAmelCase ( self : Optional[int] ) -> None: # Convert into blocks of 64 bytes __magic_name__ : str = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers __magic_name__ : Optional[int] = list(struct.unpack('>16L' , _A ) ) # add 48 0-ed integers words += [0] * 48 __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Optional[int] = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array __magic_name__ : Optional[int] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) __magic_name__ : Any = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) __magic_name__ : List[Any] = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0_0_0_0_0_0_0_0 # Compression __magic_name__ : List[str] = self.ror(_A , 6 ) ^ self.ror(_A , 11 ) ^ self.ror(_A , 25 ) __magic_name__ : Tuple = (e & f) ^ ((~e & 0XF_F_F_F_F_F_F_F) & g) __magic_name__ : Optional[int] = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0_0_0_0_0_0_0_0 __magic_name__ : Tuple = self.ror(_A , 2 ) ^ self.ror(_A , 13 ) ^ self.ror(_A , 22 ) __magic_name__ : Optional[Any] = (a & b) ^ (a & c) ^ (b & c) __magic_name__ : Optional[int] = (sa + maj) % 0X1_0_0_0_0_0_0_0_0 __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : str = ( g, f, e, ((d + tempa) % 0X1_0_0_0_0_0_0_0_0), c, b, a, ((tempa + tempa) % 0X1_0_0_0_0_0_0_0_0), ) __magic_name__ : Tuple = [a, b, c, d, e, f, g, h] # Modify final values __magic_name__ : List[str] = [ ((element + mutated_hash_values[index]) % 0X1_0_0_0_0_0_0_0_0) for index, element in enumerate(self.hashes ) ] __magic_name__ : List[str] = ''.join([hex(_A )[2:].zfill(8 ) for value in self.hashes] ) def __lowerCAmelCase ( self : int , _A : int , _A : int ) -> int: return 0XF_F_F_F_F_F_F_F & (value << (32 - rotations)) | (value >> rotations) class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : Tuple ) -> None: import hashlib __magic_name__ : int = bytes('Test String' , 'utf-8' ) self.assertEqual(SHAaaa(_A ).hash , hashlib.shaaaa(_A ).hexdigest() ) def lowerCamelCase ( ): """simple docstring""" import doctest doctest.testmod() __magic_name__ : Optional[int] = argparse.ArgumentParser() parser.add_argument( '-s' , '--string' , dest='input_string' , default='Hello World!! Welcome to Cryptography' , help='Hash the string' , ) parser.add_argument( '-f' , '--file' , dest='input_file' , help='Hash contents of a file' ) __magic_name__ : Tuple = parser.parse_args() __magic_name__ : Optional[Any] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , 'rb' ) as f: __magic_name__ : List[Any] = f.read() else: __magic_name__ : Dict = bytes(lowerCAmelCase , 'utf-8' ) print(SHAaaa(lowerCAmelCase ).hash ) if __name__ == "__main__": main()
331
'''simple docstring''' class _lowerCamelCase : # Public class to implement a graph '''simple docstring''' def __init__( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> None: __magic_name__ : Tuple = row __magic_name__ : str = col __magic_name__ : Optional[Any] = graph def __lowerCAmelCase ( self : Any , _A : int , _A : int , _A : list[list[bool]] ) -> bool: return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def __lowerCAmelCase ( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> None: # Checking all 8 elements surrounding nth element __magic_name__ : List[str] = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __magic_name__ : List[str] = [-1, 0, 1, -1, 1, -1, 0, 1] __magic_name__ : Optional[int] = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , _A ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , _A ) def __lowerCAmelCase ( self : int ) -> int: # And finally, count all islands. __magic_name__ : List[str] = [[False for j in range(self.COL )] for i in range(self.ROW )] __magic_name__ : Any = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(_A , _A , _A ) count += 1 return count
331
1
import operator as op def A ( a_ ) -> Optional[Any]: __UpperCamelCase : str =[] __UpperCamelCase : List[str] =lambda a_ ,a_ : int(x / y ) # noqa: E731 integer division operation __UpperCamelCase : Optional[int] ={ '''^''': op.pow, '''*''': op.mul, '''/''': div, '''+''': op.add, '''-''': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) ,'Action'.center(12 ) ,'Stack' ,sep=' | ' ) print('-' * (30 + len(__UpperCAmelCase )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(__UpperCAmelCase ) # append x to stack # output in tabular format print(x.rjust(8 ) ,('push(' + x + ')').ljust(12 ) ,','.join(__UpperCAmelCase ) ,sep=' | ' ) else: __UpperCamelCase : List[Any] =stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) ,('pop(' + b + ')').ljust(12 ) ,','.join(__UpperCAmelCase ) ,sep=' | ' ) __UpperCamelCase : str =stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) ,('pop(' + a + ')').ljust(12 ) ,','.join(__UpperCAmelCase ) ,sep=' | ' ) stack.append( str(opr[x](int(__UpperCAmelCase ) ,int(__UpperCAmelCase ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) ,('push(' + a + x + b + ')').ljust(12 ) ,','.join(__UpperCAmelCase ) ,sep=' | ' ,) return int(stack[0] ) if __name__ == "__main__": A_ :List[str] = input('''\n\nEnter a Postfix Equation (space separated) = ''').split(''' ''') print('''\n\tResult = ''', solve(Postfix))
352
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def A ( a_ = 3 ) -> qiskit.result.counts.Counts: if isinstance(a_ ,a_ ): raise TypeError('number of qubits must be a integer.' ) if number_of_qubits <= 0: raise ValueError('number of qubits must be > 0.' ) if math.floor(a_ ) != number_of_qubits: raise ValueError('number of qubits must be exact integer.' ) if number_of_qubits > 10: raise ValueError('number of qubits too large to simulate(>10).' ) __UpperCamelCase : str =QuantumRegister(a_ ,'qr' ) __UpperCamelCase : Optional[int] =ClassicalRegister(a_ ,'cr' ) __UpperCamelCase : Optional[Any] =QuantumCircuit(a_ ,a_ ) __UpperCamelCase : Any =number_of_qubits for i in range(a_ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(a_ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) ,a_ ,a_ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(a_ ,number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(a_ ,a_ ) # simulate with 10000 shots __UpperCamelCase : Any =Aer.get_backend('qasm_simulator' ) __UpperCamelCase : Tuple =execute(a_ ,a_ ,shots=10_000 ) return job.result().get_counts(a_ ) if __name__ == "__main__": print( f"Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}" )
245
0
from collections import defaultdict def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> bool: """simple docstring""" A : Any = first_str.lower().strip() A : Any = second_str.lower().strip() # Remove whitespace A : Any = first_str.replace(""" """ , """""" ) A : Union[str, Any] = second_str.replace(""" """ , """""" ) # Strings of different lengths are not anagrams if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): return False # Default values for count should be 0 A : defaultdict[str, int] = defaultdict(_lowerCAmelCase ) # For each character in input strings, # increment count in the corresponding for i in range(len(_lowerCAmelCase ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() SCREAMING_SNAKE_CASE_:Union[str, Any] = input("""Enter the first string """).strip() SCREAMING_SNAKE_CASE_:Any = input("""Enter the second string """).strip() SCREAMING_SNAKE_CASE_:Optional[int] = check_anagrams(input_a, input_b) print(F"""{input_a} and {input_b} are {"" if status else "not "}anagrams.""")
116
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE_:Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_:Dict = { """facebook/xmod-base""": """https://huggingface.co/facebook/xmod-base/resolve/main/config.json""", """facebook/xmod-large-prenorm""": """https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json""", """facebook/xmod-base-13-125k""": """https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json""", """facebook/xmod-base-30-125k""": """https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json""", """facebook/xmod-base-30-195k""": """https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json""", """facebook/xmod-base-60-125k""": """https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json""", """facebook/xmod-base-60-265k""": """https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json""", """facebook/xmod-base-75-125k""": """https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json""", """facebook/xmod-base-75-269k""": """https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json""", } class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __lowerCamelCase : Dict = "xmod" def __init__( self, lowerCamelCase__=3_0522, lowerCamelCase__=768, lowerCamelCase__=12, lowerCamelCase__=12, lowerCamelCase__=3072, lowerCamelCase__="gelu", lowerCamelCase__=0.1, lowerCamelCase__=0.1, lowerCamelCase__=512, lowerCamelCase__=2, lowerCamelCase__=0.02, lowerCamelCase__=1e-12, lowerCamelCase__=1, lowerCamelCase__=0, lowerCamelCase__=2, lowerCamelCase__="absolute", lowerCamelCase__=True, lowerCamelCase__=None, lowerCamelCase__=False, lowerCamelCase__=2, lowerCamelCase__=False, lowerCamelCase__=True, lowerCamelCase__=True, lowerCamelCase__=("en_XX",), lowerCamelCase__=None, **lowerCamelCase__, ): super().__init__(pad_token_id=lowerCamelCase__, bos_token_id=lowerCamelCase__, eos_token_id=lowerCamelCase__, **lowerCamelCase__ ) A : int = vocab_size A : int = hidden_size A : str = num_hidden_layers A : List[str] = num_attention_heads A : List[str] = hidden_act A : Dict = intermediate_size A : Optional[int] = hidden_dropout_prob A : Union[str, Any] = attention_probs_dropout_prob A : int = max_position_embeddings A : Tuple = type_vocab_size A : List[Any] = initializer_range A : str = layer_norm_eps A : Union[str, Any] = position_embedding_type A : Any = use_cache A : int = classifier_dropout A : int = pre_norm A : List[str] = adapter_reduction_factor A : Any = adapter_layer_norm A : Any = adapter_reuse_layer_norm A : str = ln_before_adapter A : Dict = list(lowerCamelCase__ ) A : Union[str, Any] = default_language class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' @property def _lowerCAmelCase ( self ): if self.task == "multiple-choice": A : Any = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
116
1
import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class a_ : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=6 , _lowerCamelCase=17 , _lowerCamelCase=23 , _lowerCamelCase=11 , _lowerCamelCase=True , ) ->List[Any]: SCREAMING_SNAKE_CASE : List[Any] = parent SCREAMING_SNAKE_CASE : Optional[Any] = batch_size SCREAMING_SNAKE_CASE : List[str] = seq_length SCREAMING_SNAKE_CASE : List[str] = act_dim SCREAMING_SNAKE_CASE : Tuple = state_dim SCREAMING_SNAKE_CASE : List[Any] = hidden_size SCREAMING_SNAKE_CASE : Dict = max_length SCREAMING_SNAKE_CASE : List[str] = is_training def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : int = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) SCREAMING_SNAKE_CASE : int = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) SCREAMING_SNAKE_CASE : List[Any] = floats_tensor((self.batch_size, self.seq_length, 1) ) SCREAMING_SNAKE_CASE : Union[str, Any] = floats_tensor((self.batch_size, self.seq_length, 1) ) SCREAMING_SNAKE_CASE : Tuple = ids_tensor((self.batch_size, self.seq_length) , vocab_size=1000 ) SCREAMING_SNAKE_CASE : List[str] = random_attention_mask((self.batch_size, self.seq_length) ) SCREAMING_SNAKE_CASE : Any = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def __lowerCAmelCase ( self ) ->Any: return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) ->Optional[int]: SCREAMING_SNAKE_CASE : List[Any] = DecisionTransformerModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE : Optional[int] = model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Tuple = self.prepare_config_and_inputs() ( SCREAMING_SNAKE_CASE ) : Dict = config_and_inputs SCREAMING_SNAKE_CASE : Union[str, Any] = { '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class a_ ( a__ , a__ , a__ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = (DecisionTransformerModel,) if is_torch_available() else () __SCREAMING_SNAKE_CASE : Tuple = () __SCREAMING_SNAKE_CASE : str = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids __SCREAMING_SNAKE_CASE : List[str] = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features __SCREAMING_SNAKE_CASE : int = False __SCREAMING_SNAKE_CASE : Optional[int] = False __SCREAMING_SNAKE_CASE : Optional[int] = False __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : str = False __SCREAMING_SNAKE_CASE : int = False __SCREAMING_SNAKE_CASE : List[str] = False __SCREAMING_SNAKE_CASE : List[Any] = False def __lowerCAmelCase ( self ) ->Dict: SCREAMING_SNAKE_CASE : List[str] = DecisionTransformerModelTester(self ) SCREAMING_SNAKE_CASE : Union[str, Any] = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ) ->List[Any]: self.config_tester.run_common_tests() def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) @slow def __lowerCAmelCase ( self ) ->Optional[Any]: for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE : Dict = DecisionTransformerModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Any = model_class(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE : int = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE : Any = [ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(_lowerCamelCase )] , _lowerCamelCase ) @require_torch class a_ ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : str = 2 # number of steps of autoregressive prediction we will perform SCREAMING_SNAKE_CASE : List[Any] = 10 # defined by the RL environment, may be normalized SCREAMING_SNAKE_CASE : int = DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) SCREAMING_SNAKE_CASE : List[Any] = model.to(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = model.config torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : str = torch.randn(1 , 1 , config.state_dim ).to(device=_lowerCamelCase , dtype=torch.floataa ) # env.reset() SCREAMING_SNAKE_CASE : Optional[int] = torch.tensor( [[0.2_4_2_7_9_3, -0.2_8_6_9_3_0_7_4, 0.8_7_4_2_6_1_3], [0.6_7_8_1_5_2_7_4, -0.0_8_1_0_1_0_8_5, -0.1_2_9_5_2_1_4_7]] , device=_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = torch.tensor(_lowerCamelCase , device=_lowerCamelCase , dtype=torch.floataa ).reshape(1 , 1 , 1 ) SCREAMING_SNAKE_CASE : List[str] = state SCREAMING_SNAKE_CASE : Optional[int] = torch.zeros(1 , 0 , config.act_dim , device=_lowerCamelCase , dtype=torch.floataa ) SCREAMING_SNAKE_CASE : Union[str, Any] = torch.zeros(1 , 0 , device=_lowerCamelCase , dtype=torch.floataa ) SCREAMING_SNAKE_CASE : str = torch.tensor(0 , device=_lowerCamelCase , dtype=torch.long ).reshape(1 , 1 ) for step in range(_lowerCamelCase ): SCREAMING_SNAKE_CASE : Any = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=_lowerCamelCase )] , dim=1 ) SCREAMING_SNAKE_CASE : str = torch.cat([rewards, torch.zeros(1 , 1 , device=_lowerCamelCase )] , dim=1 ) SCREAMING_SNAKE_CASE : Tuple = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): SCREAMING_SNAKE_CASE : int = model( states=_lowerCamelCase , actions=_lowerCamelCase , rewards=_lowerCamelCase , returns_to_go=_lowerCamelCase , timesteps=_lowerCamelCase , attention_mask=_lowerCamelCase , return_dict=_lowerCamelCase , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1e-4 ) ) SCREAMING_SNAKE_CASE : Optional[int] = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=_lowerCamelCase , dtype=torch.floataa ), 1.0, False, {}, ) SCREAMING_SNAKE_CASE : Dict = action_pred[0, -1] SCREAMING_SNAKE_CASE : Optional[Any] = torch.cat([states, state] , dim=1 ) SCREAMING_SNAKE_CASE : Tuple = returns_to_go[0, -1] - reward SCREAMING_SNAKE_CASE : Any = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) SCREAMING_SNAKE_CASE : Optional[Any] = torch.cat( [timesteps, torch.ones((1, 1) , device=_lowerCamelCase , dtype=torch.long ) * (step + 1)] , dim=1 )
365
import math a__ : List[str] = 10 a__ : Optional[int] = 7 a__ : int = BALLS_PER_COLOUR * NUM_COLOURS def UpperCAmelCase_( a__ = 20 ): """simple docstring""" SCREAMING_SNAKE_CASE : str = math.comb(a__ , a__ ) SCREAMING_SNAKE_CASE : Dict = math.comb(NUM_BALLS - BALLS_PER_COLOUR , a__ ) SCREAMING_SNAKE_CASE : Any = NUM_COLOURS * (1 - missing_colour / total) return F"""{result:.9f}""" if __name__ == "__main__": print(solution(20))
19
0
"""simple docstring""" import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase__ ( lowerCAmelCase__ :int , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Tuple ) -> Optional[int]: '''simple docstring''' lowercase = RemBertConfig.from_json_file(lowerCAmelCase__ ) print("""Building PyTorch model from configuration: {}""".format(str(lowerCAmelCase__ ) ) ) lowercase = RemBertModel(lowerCAmelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_rembert(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Save pytorch-model print("""Save PyTorch model to {}""".format(lowerCAmelCase__ ) ) torch.save(model.state_dict() , lowerCAmelCase__ ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--rembert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained RemBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase : Union[str, Any] =parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
197
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging a_ : Optional[int] = logging.get_logger(__name__) a_ : Dict = { "Helsinki-NLP/opus-mt-en-de": "https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json", # See all Marian models at https://huggingface.co/models?filter=marian } class a ( _SCREAMING_SNAKE_CASE ): _lowerCAmelCase = """marian""" _lowerCAmelCase = ["""past_key_values"""] _lowerCAmelCase = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , __magic_name__=5_81_01 , __magic_name__=None , __magic_name__=10_24 , __magic_name__=12 , __magic_name__=40_96 , __magic_name__=16 , __magic_name__=12 , __magic_name__=40_96 , __magic_name__=16 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__=True , __magic_name__="gelu" , __magic_name__=10_24 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0_2 , __magic_name__=5_81_00 , __magic_name__=False , __magic_name__=5_81_00 , __magic_name__=0 , __magic_name__=0 , __magic_name__=True , **__magic_name__ , ) -> str: _a = vocab_size _a = decoder_vocab_size or vocab_size _a = max_position_embeddings _a = d_model _a = encoder_ffn_dim _a = encoder_layers _a = encoder_attention_heads _a = decoder_ffn_dim _a = decoder_layers _a = decoder_attention_heads _a = dropout _a = attention_dropout _a = activation_dropout _a = activation_function _a = init_std _a = encoder_layerdrop _a = decoder_layerdrop _a = use_cache _a = encoder_layers _a = scale_embedding # scale factor will be sqrt(d_model) if True _a = share_encoder_decoder_embeddings super().__init__( pad_token_id=__magic_name__ , eos_token_id=__magic_name__ , is_encoder_decoder=__magic_name__ , decoder_start_token_id=__magic_name__ , forced_eos_token_id=__magic_name__ , **__magic_name__ , ) class a ( _SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def __UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _a = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: _a = {0: 'batch'} _a = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: _a = {0: 'batch', 1: 'decoder_sequence'} _a = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. _a = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: _a , _a = self.num_layers for i in range(__magic_name__ ): _a = {0: 'batch', 2: 'past_sequence + sequence'} _a = {0: 'batch', 2: 'past_sequence + sequence'} else: _a = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def __UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _a = super().outputs else: _a = super(__magic_name__ , self ).outputs if self.use_past: _a , _a = self.num_layers for i in range(__magic_name__ ): _a = {0: 'batch', 2: 'past_sequence + sequence'} _a = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ) -> Mapping[str, Any]: _a = self._generate_dummy_inputs_for_encoder_and_decoder( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Generate decoder inputs _a = seq_length if not self.use_past else 1 _a = self._generate_dummy_inputs_for_encoder_and_decoder( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) _a = {f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} _a = dict(**__magic_name__ , **__magic_name__ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _a , _a = common_inputs['input_ids'].shape _a = common_inputs['decoder_input_ids'].shape[1] _a , _a = self.num_attention_heads _a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) _a = decoder_seq_length + 3 _a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) _a = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(__magic_name__ , __magic_name__ )] , dim=1 ) _a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered _a , _a = self.num_layers _a = min(__magic_name__ , __magic_name__ ) _a = max(__magic_name__ , __magic_name__ ) - min_num_layers _a = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(__magic_name__ ): common_inputs["past_key_values"].append( ( torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ ), ) ) # TODO: test this. _a = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(__magic_name__ , __magic_name__ ): common_inputs["past_key_values"].append((torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) ) return common_inputs def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ) -> Mapping[str, Any]: _a = self._generate_dummy_inputs_for_encoder_and_decoder( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _a , _a = common_inputs['input_ids'].shape # Not using the same length for past_key_values _a = seqlen + 2 _a , _a = self.num_layers _a , _a = self.num_attention_heads _a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) _a = common_inputs['attention_mask'].dtype _a = torch.cat( [common_inputs['attention_mask'], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) _a = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(__magic_name__ ) ] return common_inputs def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _a = compute_effective_axis_dimension( __magic_name__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _a = tokenizer.num_special_tokens_to_add(__magic_name__ ) _a = compute_effective_axis_dimension( __magic_name__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__magic_name__ ) # Generate dummy inputs according to compute batch and sequence _a = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size _a = dict(tokenizer(__magic_name__ , return_tensors=__magic_name__ ) ) return common_inputs def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: _a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) else: _a = self._generate_dummy_inputs_for_causal_lm( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) return common_inputs def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Tuple: if self.task in ["default", "seq2seq-lm"]: _a = super()._flatten_past_key_values_(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: _a = super(__magic_name__ , self )._flatten_past_key_values_( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) @property def __UpperCAmelCase ( self ) -> float: return 1e-4
168
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case : Tuple = { 'configuration_blenderbot_small': [ 'BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotSmallConfig', 'BlenderbotSmallOnnxConfig', ], 'tokenization_blenderbot_small': ['BlenderbotSmallTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = ['BlenderbotSmallTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotSmallForCausalLM', 'BlenderbotSmallForConditionalGeneration', 'BlenderbotSmallModel', 'BlenderbotSmallPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = [ 'TFBlenderbotSmallForConditionalGeneration', 'TFBlenderbotSmallModel', 'TFBlenderbotSmallPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Dict = [ 'FlaxBlenderbotSmallForConditionalGeneration', 'FlaxBlenderbotSmallModel', 'FlaxBlenderbotSmallPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
207
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def a_ ( lowerCAmelCase_ : Dict[str, torch.Tensor] ): __lowerCAmelCase = [] __lowerCAmelCase = [] __lowerCAmelCase = [] for rt in rc.restypes: __lowerCAmelCase = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) __lowerCAmelCase = {name: i for i, name in enumerate(lowerCAmelCase_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) __lowerCAmelCase = torch.tensor( lowerCAmelCase_, dtype=torch.intaa, device=protein['aatype'].device, ) __lowerCAmelCase = torch.tensor( lowerCAmelCase_, dtype=torch.intaa, device=protein['aatype'].device, ) __lowerCAmelCase = torch.tensor( lowerCAmelCase_, dtype=torch.floataa, device=protein['aatype'].device, ) __lowerCAmelCase = protein['aatype'].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein __lowerCAmelCase = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase = restype_atomaa_mask[protein_aatype] __lowerCAmelCase = residx_atomaa_mask __lowerCAmelCase = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back __lowerCAmelCase = restype_atomaa_to_atomaa[protein_aatype] __lowerCAmelCase = residx_atomaa_to_atomaa.long() # create the corresponding mask __lowerCAmelCase = torch.zeros([21, 37], dtype=torch.floataa, device=protein['aatype'].device ) for restype, restype_letter in enumerate(rc.restypes ): __lowerCAmelCase = rc.restype_atoa[restype_letter] __lowerCAmelCase = rc.residue_atoms[restype_name] for atom_name in atom_names: __lowerCAmelCase = rc.atom_order[atom_name] __lowerCAmelCase = 1 __lowerCAmelCase = restype_atomaa_mask[protein_aatype] __lowerCAmelCase = residx_atomaa_mask return protein def a_ ( lowerCAmelCase_ : Dict[str, torch.Tensor] ): __lowerCAmelCase = tree_map(lambda lowerCAmelCase_ : torch.tensor(lowerCAmelCase_, device=batch['aatype'].device ), lowerCAmelCase_, np.ndarray ) __lowerCAmelCase = tensor_tree_map(lambda lowerCAmelCase_ : np.array(lowerCAmelCase_ ), make_atomaa_masks(lowerCAmelCase_ ) ) return out
207
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A : Any = logging.get_logger(__name__) __A : str = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __A ( lowerCAmelCase ): lowerCAmelCase_ : Optional[int] = "distilbert" lowerCAmelCase_ : int = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self : List[Any] , UpperCAmelCase_ : Optional[Any]=30522 , UpperCAmelCase_ : Tuple=512 , UpperCAmelCase_ : List[str]=False , UpperCAmelCase_ : Optional[Any]=6 , UpperCAmelCase_ : Tuple=12 , UpperCAmelCase_ : List[str]=768 , UpperCAmelCase_ : List[str]=4 * 768 , UpperCAmelCase_ : Union[str, Any]=0.1 , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : int="gelu" , UpperCAmelCase_ : int=0.02 , UpperCAmelCase_ : str=0.1 , UpperCAmelCase_ : Dict=0.2 , UpperCAmelCase_ : Union[str, Any]=0 , **UpperCAmelCase_ : str , ): lowerCAmelCase : Tuple = vocab_size lowerCAmelCase : Optional[Any] = max_position_embeddings lowerCAmelCase : int = sinusoidal_pos_embds lowerCAmelCase : List[str] = n_layers lowerCAmelCase : Dict = n_heads lowerCAmelCase : int = dim lowerCAmelCase : Optional[Any] = hidden_dim lowerCAmelCase : str = dropout lowerCAmelCase : List[str] = attention_dropout lowerCAmelCase : Optional[int] = activation lowerCAmelCase : int = initializer_range lowerCAmelCase : Optional[int] = qa_dropout lowerCAmelCase : Any = seq_classif_dropout super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ ) class __A ( lowerCAmelCase ): @property def lowercase__ ( self : Optional[Any] ): if self.task == "multiple-choice": lowerCAmelCase : int = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCAmelCase : Optional[int] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
138
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class __A : @staticmethod def lowercase__ ( *UpperCAmelCase_ : Tuple , **UpperCAmelCase_ : Any ): pass def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> str: '''simple docstring''' lowerCAmelCase : List[Any] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Dict: '''simple docstring''' lowerCAmelCase : Tuple = np.array(_UpperCAmelCase ) lowerCAmelCase : Dict = npimg.shape return {"hash": hashimage(_UpperCAmelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class __A ( unittest.TestCase ): lowerCAmelCase_ : Dict = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) lowerCAmelCase_ : Any = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[str] ): lowerCAmelCase : List[str] = MaskGenerationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def lowercase__ ( self : Tuple , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Dict ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def lowercase__ ( self : Dict ): pass @slow @require_torch def lowercase__ ( self : str ): lowerCAmelCase : Optional[int] = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) lowerCAmelCase : Union[str, Any] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=256 ) # Shortening by hashing lowerCAmelCase : List[str] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(UpperCAmelCase_ ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.04_44}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_21}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.01_67}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.01_32}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.00_53}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.99_67}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_93}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.99_09}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.98_79}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.98_34}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.97_16}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.96_12}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.95_99}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.95_52}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.95_32}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.95_16}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.94_99}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.94_83}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.94_64}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_43}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_43}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.94_08}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.93_35}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.93_26}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.92_62}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.89_99}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.89_86}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.89_84}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.88_73}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.88_71} ] , ) # fmt: on @require_torch @slow def lowercase__ ( self : List[Any] ): lowerCAmelCase : Union[str, Any] = 'facebook/sam-vit-huge' lowerCAmelCase : str = pipeline('mask-generation' , model=UpperCAmelCase_ ) lowerCAmelCase : int = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing lowerCAmelCase : Optional[int] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(UpperCAmelCase_ ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.04_44}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.02_10}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.01_67}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.01_32}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.00_53}, ] , )
138
1
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class __magic_name__ ( _a): def __init__( self : List[str] , lowerCamelCase__ : str = "▁" , lowerCamelCase__ : bool = True , lowerCamelCase__ : Union[str, AddedToken] = "<unk>" , lowerCamelCase__ : Union[str, AddedToken] = "</s>" , lowerCamelCase__ : Union[str, AddedToken] = "<pad>" , ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : List[str] = { """pad""": {"""id""": 0, """token""": pad_token}, """eos""": {"""id""": 1, """token""": eos_token}, """unk""": {"""id""": 2, """token""": unk_token}, } UpperCamelCase__ : Dict = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): UpperCamelCase__ : List[str] = token_dict["""token"""] UpperCamelCase__ : List[Any] = Tokenizer(Unigram() ) UpperCamelCase__ : Optional[int] = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(''' {2,}''' ) , ''' ''' ), normalizers.Lowercase(), ] ) UpperCamelCase__ : Optional[Any] = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=snake_case_ , add_prefix_space=snake_case_ ), pre_tokenizers.Digits(individual_digits=snake_case_ ), pre_tokenizers.Punctuation(), ] ) UpperCamelCase__ : Dict = decoders.Metaspace(replacement=snake_case_ , add_prefix_space=snake_case_ ) UpperCamelCase__ : str = TemplateProcessing( single=F"$A {self.special_tokens['eos']['token']}" , special_tokens=[(self.special_tokens['''eos''']['''token'''], self.special_tokens['''eos''']['''id'''])] , ) UpperCamelCase__ : Tuple = { """model""": """SentencePieceUnigram""", """replacement""": replacement, """add_prefix_space""": add_prefix_space, } super().__init__(snake_case_ , snake_case_ ) def UpperCAmelCase__ ( self : int , lowerCamelCase__ : Union[str, List[str]] , lowerCamelCase__ : int = 8000 , lowerCamelCase__ : bool = True , ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = trainers.UnigramTrainer( vocab_size=snake_case_ , special_tokens=self.special_tokens_list , show_progress=snake_case_ , ) if isinstance(snake_case_ , snake_case_ ): UpperCamelCase__ : List[str] = [files] self._tokenizer.train(snake_case_ , trainer=snake_case_ ) self.add_unk_id() def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : Union[Iterator[str], Iterator[Iterator[str]]] , lowerCamelCase__ : int = 8000 , lowerCamelCase__ : bool = True , ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[str] = trainers.UnigramTrainer( vocab_size=snake_case_ , special_tokens=self.special_tokens_list , show_progress=snake_case_ , ) self._tokenizer.train_from_iterator(snake_case_ , trainer=snake_case_ ) self.add_unk_id() def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' UpperCamelCase__ : Any = json.loads(self._tokenizer.to_str() ) UpperCamelCase__ : Optional[int] = self.special_tokens["""unk"""]["""id"""] UpperCamelCase__ : Dict = Tokenizer.from_str(json.dumps(snake_case_ ) )
371
def _a ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : int ): """simple docstring""" if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_to_repay <= 0 or not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise Exception('''Years to repay must be an integer > 0''' ) # Yearly rate is divided by 12 to get monthly rate UpperCamelCase__ : int = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly UpperCamelCase__ : int = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
51
0
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = ["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ = "BlipImageProcessor" SCREAMING_SNAKE_CASE_ = ("BertTokenizer", "BertTokenizerFast") def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> List[Any]: snake_case_ = False super().__init__(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = self.image_processor def __call__( self, lowerCAmelCase__ = None, lowerCAmelCase__ = None, lowerCAmelCase__ = True, lowerCAmelCase__ = False, lowerCAmelCase__ = None, lowerCAmelCase__ = None, lowerCAmelCase__ = 0, lowerCAmelCase__ = None, lowerCAmelCase__ = None, lowerCAmelCase__ = False, lowerCAmelCase__ = False, lowerCAmelCase__ = False, lowerCAmelCase__ = False, lowerCAmelCase__ = False, lowerCAmelCase__ = True, lowerCAmelCase__ = None, **lowerCAmelCase__, ) -> BatchEncoding: if images is None and text is None: raise ValueError('You have to specify either images or text.') # Get only text if images is None: snake_case_ = self.tokenizer snake_case_ = self.tokenizer( text=lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, padding=lowerCAmelCase__, truncation=lowerCAmelCase__, max_length=lowerCAmelCase__, stride=lowerCAmelCase__, pad_to_multiple_of=lowerCAmelCase__, return_attention_mask=lowerCAmelCase__, return_overflowing_tokens=lowerCAmelCase__, return_special_tokens_mask=lowerCAmelCase__, return_offsets_mapping=lowerCAmelCase__, return_token_type_ids=lowerCAmelCase__, return_length=lowerCAmelCase__, verbose=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__, ) return text_encoding # add pixel_values snake_case_ = self.image_processor(lowerCAmelCase__, return_tensors=lowerCAmelCase__) if text is not None: snake_case_ = self.tokenizer( text=lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, padding=lowerCAmelCase__, truncation=lowerCAmelCase__, max_length=lowerCAmelCase__, stride=lowerCAmelCase__, pad_to_multiple_of=lowerCAmelCase__, return_attention_mask=lowerCAmelCase__, return_overflowing_tokens=lowerCAmelCase__, return_special_tokens_mask=lowerCAmelCase__, return_offsets_mapping=lowerCAmelCase__, return_token_type_ids=lowerCAmelCase__, return_length=lowerCAmelCase__, verbose=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__, ) else: snake_case_ = None if text_encoding is not None: encoding_image_processor.update(lowerCAmelCase__) return encoding_image_processor def a_ ( self, *lowerCAmelCase__, **lowerCAmelCase__) -> Optional[Any]: return self.tokenizer.batch_decode(*lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self, *lowerCAmelCase__, **lowerCAmelCase__) -> Union[str, Any]: return self.tokenizer.decode(*lowerCAmelCase__, **lowerCAmelCase__) @property def a_ ( self) -> List[str]: snake_case_ = self.tokenizer.model_input_names snake_case_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
69
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # Base Case if curr_ind == len(UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(UpperCAmelCase ) ): if valid_connection(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # Insert current vertex into path as next transition snake_case_ = next_ver # Validate created path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , curr_ind + 1 ): return True # Backtrack snake_case_ = -1 return False def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = 0 ) -> list[int]: snake_case_ = [-1] * (len(UpperCAmelCase ) + 1) # initialize start and end of path with starting index snake_case_ = snake_case_ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , 1 ) else []
69
1
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] ) -> str: UpperCAmelCase_ = '' for word_or_phrase in separated: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise Exception('''join() accepts only strings to be joined''' ) joined += word_or_phrase + separator return joined.strip(lowerCAmelCase_ ) if __name__ == "__main__": from doctest import testmod testmod()
369
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : float , __UpperCamelCase : float ) -> float: if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
177
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowercase = logging.get_logger(__name__) def __UpperCAmelCase ( a_ , a_=False): snake_case_ = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('head'): snake_case_ = 'segformer.encoder.' + key if key.startswith('backbone'): snake_case_ = key.replace('backbone' , 'segformer.encoder') if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 snake_case_ = key[key.find('patch_embed') + len('patch_embed')] snake_case_ = key.replace(f'''patch_embed{idx}''' , f'''patch_embeddings.{int(a_)-1}''') if "norm" in key: snake_case_ = key.replace('norm' , 'layer_norm') if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 snake_case_ = key[key.find('segformer.encoder.layer_norm') + len('segformer.encoder.layer_norm')] snake_case_ = key.replace(f'''layer_norm{idx}''' , f'''layer_norm.{int(a_)-1}''') if "layer_norm1" in key: snake_case_ = key.replace('layer_norm1' , 'layer_norm_1') if "layer_norm2" in key: snake_case_ = key.replace('layer_norm2' , 'layer_norm_2') if "block" in key: # replace for example block1 by block.0 snake_case_ = key[key.find('block') + len('block')] snake_case_ = key.replace(f'''block{idx}''' , f'''block.{int(a_)-1}''') if "attn.q" in key: snake_case_ = key.replace('attn.q' , 'attention.self.query') if "attn.proj" in key: snake_case_ = key.replace('attn.proj' , 'attention.output.dense') if "attn" in key: snake_case_ = key.replace('attn' , 'attention.self') if "fc1" in key: snake_case_ = key.replace('fc1' , 'dense1') if "fc2" in key: snake_case_ = key.replace('fc2' , 'dense2') if "linear_pred" in key: snake_case_ = key.replace('linear_pred' , 'classifier') if "linear_fuse" in key: snake_case_ = key.replace('linear_fuse.conv' , 'linear_fuse') snake_case_ = key.replace('linear_fuse.bn' , 'batch_norm') if "linear_c" in key: # replace for example linear_c4 by linear_c.3 snake_case_ = key[key.find('linear_c') + len('linear_c')] snake_case_ = key.replace(f'''linear_c{idx}''' , f'''linear_c.{int(a_)-1}''') if key.startswith('head'): snake_case_ = key.replace('head' , 'classifier') snake_case_ = value return new_state_dict def __UpperCAmelCase ( a_ , a_): # for each of the encoder blocks: for i in range(config.num_encoder_blocks): for j in range(config.depths[i]): # read in weights + bias of keys and values (which is a single matrix in the original implementation) snake_case_ = state_dict.pop(f'''segformer.encoder.block.{i}.{j}.attention.self.kv.weight''') snake_case_ = state_dict.pop(f'''segformer.encoder.block.{i}.{j}.attention.self.kv.bias''') # next, add keys and values (in that order) to the state dict snake_case_ = kv_weight[ : config.hidden_sizes[i], : ] snake_case_ = kv_bias[: config.hidden_sizes[i]] snake_case_ = kv_weight[ config.hidden_sizes[i] :, : ] snake_case_ = kv_bias[ config.hidden_sizes[i] : ] def __UpperCAmelCase ( ): snake_case_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' snake_case_ = Image.open(requests.get(a_ , stream=a_).raw) return image @torch.no_grad() def __UpperCAmelCase ( a_ , a_ , a_): snake_case_ = SegformerConfig() snake_case_ = False # set attributes based on model_name snake_case_ = 'huggingface/label-files' if "segformer" in model_name: snake_case_ = model_name[len('segformer.') : len('segformer.') + 2] if "ade" in model_name: snake_case_ = 1_50 snake_case_ = 'ade20k-id2label.json' snake_case_ = (1, 1_50, 1_28, 1_28) elif "city" in model_name: snake_case_ = 19 snake_case_ = 'cityscapes-id2label.json' snake_case_ = (1, 19, 1_28, 1_28) else: raise ValueError(f'''Model {model_name} not supported''') elif "mit" in model_name: snake_case_ = True snake_case_ = model_name[4:6] snake_case_ = 10_00 snake_case_ = 'imagenet-1k-id2label.json' snake_case_ = (1, 10_00) else: raise ValueError(f'''Model {model_name} not supported''') # set config attributes snake_case_ = json.load(open(hf_hub_download(a_ , a_ , repo_type='dataset') , 'r')) snake_case_ = {int(a_): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": snake_case_ = [64, 1_28, 3_20, 5_12] snake_case_ = 2_56 elif size == "b2": snake_case_ = [64, 1_28, 3_20, 5_12] snake_case_ = 7_68 snake_case_ = [3, 4, 6, 3] elif size == "b3": snake_case_ = [64, 1_28, 3_20, 5_12] snake_case_ = 7_68 snake_case_ = [3, 4, 18, 3] elif size == "b4": snake_case_ = [64, 1_28, 3_20, 5_12] snake_case_ = 7_68 snake_case_ = [3, 8, 27, 3] elif size == "b5": snake_case_ = [64, 1_28, 3_20, 5_12] snake_case_ = 7_68 snake_case_ = [3, 6, 40, 3] else: raise ValueError(f'''Size {size} not supported''') # load image processor (only resize + normalize) snake_case_ = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=a_ , align=a_ , do_random_crop=a_) # prepare image snake_case_ = prepare_img() snake_case_ = image_processor(images=a_ , return_tensors='pt').pixel_values logger.info(f'''Converting model {model_name}...''') # load original state dict if encoder_only: snake_case_ = torch.load(a_ , map_location=torch.device('cpu')) else: snake_case_ = torch.load(a_ , map_location=torch.device('cpu'))['state_dict'] # rename keys snake_case_ = rename_keys(a_ , encoder_only=a_) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(a_ , a_) # create HuggingFace model and load state dict if encoder_only: snake_case_ = False snake_case_ = SegformerForImageClassification(a_) else: snake_case_ = SegformerForSemanticSegmentation(a_) model.load_state_dict(a_) model.eval() # forward pass snake_case_ = model(a_) snake_case_ = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": snake_case_ = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ]) elif model_name == "segformer.b1.512x512.ade.160k": snake_case_ = torch.tensor( [ [[-7.58_20, -8.72_31, -8.32_15], [-8.06_00, -10.35_29, -10.03_04], [-7.52_08, -9.41_03, -9.62_39]], [[-12.69_18, -13.89_94, -13.71_37], [-13.31_96, -15.75_23, -15.47_89], [-12.93_43, -14.87_57, -14.96_89]], [[-11.19_11, -11.94_21, -11.32_43], [-11.33_42, -13.68_39, -13.35_81], [-10.39_09, -12.18_32, -12.48_58]], ]) elif model_name == "segformer.b2.512x512.ade.160k": snake_case_ = torch.tensor( [ [[-11.81_73, -14.38_50, -16.31_28], [-14.56_48, -16.58_04, -18.65_68], [-14.72_23, -15.73_87, -18.42_18]], [[-15.72_90, -17.91_71, -19.44_23], [-18.31_05, -19.94_48, -21.46_61], [-17.92_96, -18.64_97, -20.79_10]], [[-15.07_83, -17.03_36, -18.27_89], [-16.87_71, -18.68_70, -20.16_12], [-16.24_54, -17.14_26, -19.50_55]], ]) elif model_name == "segformer.b3.512x512.ade.160k": snake_case_ = torch.tensor( [ [[-9.08_78, -10.20_81, -10.18_91], [-9.31_44, -10.79_41, -10.98_43], [-9.22_94, -10.38_55, -10.57_04]], [[-12.23_16, -13.90_68, -13.61_02], [-12.91_61, -14.37_02, -14.32_35], [-12.52_33, -13.71_74, -13.79_32]], [[-14.62_75, -15.24_90, -14.97_27], [-14.34_00, -15.96_87, -16.28_27], [-14.14_84, -15.40_33, -15.89_37]], ]) elif model_name == "segformer.b4.512x512.ade.160k": snake_case_ = torch.tensor( [ [[-12.31_44, -13.24_47, -14.08_02], [-13.36_14, -14.58_16, -15.61_17], [-13.33_40, -14.44_33, -16.22_19]], [[-19.27_81, -20.41_28, -20.75_06], [-20.61_53, -21.65_66, -22.09_98], [-19.98_00, -21.04_30, -22.14_94]], [[-18.87_39, -19.78_04, -21.18_34], [-20.12_33, -21.67_65, -23.29_44], [-20.03_15, -21.26_41, -23.69_44]], ]) elif model_name == "segformer.b5.640x640.ade.160k": snake_case_ = torch.tensor( [ [[-9.55_24, -12.08_35, -11.73_48], [-10.52_29, -13.64_46, -14.56_62], [-9.58_42, -12.88_51, -13.94_14]], [[-15.34_32, -17.53_23, -17.08_18], [-16.33_30, -18.92_55, -19.21_01], [-15.13_40, -17.78_48, -18.39_71]], [[-12.60_72, -14.94_86, -14.66_31], [-13.76_29, -17.09_07, -17.77_45], [-12.78_99, -16.16_95, -17.16_71]], ]) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-11.92_95, -13.40_57, -14.81_06], [-13.34_31, -14.81_79, -15.37_81], [-14.28_36, -15.59_42, -16.15_88]], [[-11.49_06, -12.80_67, -13.65_64], [-13.11_89, -14.05_00, -14.15_43], [-13.87_48, -14.51_36, -14.87_89]], [[0.53_74, 0.10_67, -0.47_42], [0.11_41, -0.22_55, -0.70_99], [-0.30_00, -0.59_24, -1.31_05]], ]) elif model_name == "segformer.b0.512x1024.city.160k": snake_case_ = torch.tensor( [ [[-7.82_17, -9.87_67, -10.17_17], [-9.44_38, -10.90_58, -11.40_47], [-9.79_39, -12.34_95, -12.10_79]], [[-7.15_14, -9.53_36, -10.08_60], [-9.77_76, -11.68_22, -11.84_39], [-10.14_11, -12.76_55, -12.89_72]], [[0.30_21, 0.08_05, -0.23_10], [-0.03_28, -0.16_05, -0.27_14], [-0.14_08, -0.54_77, -0.69_76]], ]) elif model_name == "segformer.b0.640x1280.city.160k": snake_case_ = torch.tensor( [ [ [-1.1_372E01, -1.2_787E01, -1.3_477E01], [-1.2_536E01, -1.4_194E01, -1.4_409E01], [-1.3_217E01, -1.4_888E01, -1.5_327E01], ], [ [-1.4_791E01, -1.7_122E01, -1.8_277E01], [-1.7_163E01, -1.9_192E01, -1.9_533E01], [-1.7_897E01, -1.9_991E01, -2.0_315E01], ], [ [7.6_723E-01, 4.1_921E-01, -7.7_878E-02], [4.7_772E-01, 9.5_557E-03, -2.8_082E-01], [3.6_032E-01, -2.4_826E-01, -5.1_168E-01], ], ]) elif model_name == "segformer.b0.768x768.city.160k": snake_case_ = torch.tensor( [ [[-9.49_59, -11.30_87, -11.74_79], [-11.00_25, -12.65_40, -12.33_19], [-11.40_64, -13.04_87, -12.99_05]], [[-9.89_05, -11.30_84, -12.08_54], [-11.17_26, -12.76_98, -12.95_83], [-11.59_85, -13.32_78, -14.17_74]], [[0.22_13, 0.01_92, -0.24_66], [-0.17_31, -0.42_13, -0.48_74], [-0.31_26, -0.65_41, -1.13_89]], ]) elif model_name == "segformer.b1.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ]) elif model_name == "segformer.b2.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-16.09_76, -16.48_56, -17.39_62], [-16.62_34, -19.03_42, -19.76_85], [-16.09_00, -18.06_61, -19.11_80]], [[-18.47_50, -18.84_88, -19.50_74], [-19.40_30, -22.15_70, -22.59_77], [-19.11_91, -20.84_86, -22.37_83]], [[-4.51_78, -5.50_37, -6.51_09], [-5.08_84, -7.21_74, -8.03_34], [-4.41_56, -5.81_17, -7.29_70]], ]) elif model_name == "segformer.b3.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-14.20_81, -14.47_32, -14.19_77], [-14.58_67, -16.44_23, -16.63_56], [-13.44_41, -14.96_85, -16.86_96]], [[-14.45_76, -14.70_73, -15.04_51], [-15.08_16, -17.62_37, -17.98_73], [-14.42_13, -16.01_99, -18.59_92]], [[-4.73_49, -4.95_88, -5.09_66], [-4.32_10, -6.93_25, -7.25_91], [-3.43_12, -4.74_84, -7.19_17]], ]) elif model_name == "segformer.b4.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-11.77_37, -11.95_26, -11.32_73], [-13.66_92, -14.45_74, -13.88_78], [-13.89_37, -14.69_24, -15.93_45]], [[-14.67_06, -14.53_30, -14.13_06], [-16.15_02, -16.81_80, -16.42_69], [-16.83_38, -17.89_39, -20.17_46]], [[1.04_91, 0.82_89, 1.03_10], [1.10_44, 0.52_19, 0.80_55], [1.08_99, 0.69_26, 0.55_90]], ]) elif model_name == "segformer.b5.1024x1024.city.160k": snake_case_ = torch.tensor( [ [[-12.56_41, -13.47_77, -13.06_84], [-13.95_87, -15.89_83, -16.65_57], [-13.31_09, -15.73_50, -16.31_41]], [[-14.70_74, -15.43_52, -14.59_44], [-16.63_53, -18.16_63, -18.61_20], [-15.17_02, -18.03_29, -18.15_47]], [[-1.79_90, -2.09_51, -1.77_84], [-2.63_97, -3.82_45, -3.96_86], [-1.52_64, -2.81_26, -2.93_16]], ]) else: snake_case_ = logits.argmax(-1).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx]) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , a_ , atol=1E-2) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''') Path(a_).mkdir(exist_ok=a_) model.save_pretrained(a_) image_processor.save_pretrained(a_) if __name__ == "__main__": lowercase = argparse.ArgumentParser() parser.add_argument( "--model_name", default="segformer.b0.512x512.ade.160k", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file)." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) lowercase = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
178
from __future__ import annotations def __UpperCAmelCase ( a_ , a_ , a_ , a_): # noqa: E741 while r - l > 1: snake_case_ = (l + r) // 2 if v[m] >= key: snake_case_ = m else: snake_case_ = m # noqa: E741 return r def __UpperCAmelCase ( a_): if len(a_) == 0: return 0 snake_case_ = [0] * len(a_) snake_case_ = 1 snake_case_ = v[0] for i in range(1 , len(a_)): if v[i] < tail[0]: snake_case_ = v[i] elif v[i] > tail[length - 1]: snake_case_ = v[i] length += 1 else: snake_case_ = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
178
1
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def A_ ( ) -> tuple[list[int], int]: a__ : Tuple = [randint(-1000 , 1000 ) for i in range(10 )] a__ : Dict = randint(-5000 , 5000 ) return (arr, r) lowercase : Dict = make_dataset() def A_ ( A__ , A__ ) -> tuple[int, ...]: for triplet in permutations(A__ , 3 ): if sum(A__ ) == target: return tuple(sorted(A__ ) ) return (0, 0, 0) def A_ ( A__ , A__ ) -> tuple[int, int, int]: arr.sort() a__ : Optional[Any] = len(A__ ) for i in range(n - 1 ): a__ : Dict = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def A_ ( ) -> tuple[float, float]: a__ : int = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' a__ : List[Any] = '\ntriplet_sum1(*dataset)\n' a__ : Optional[Any] = '\ntriplet_sum2(*dataset)\n' a__ : Union[str, Any] = repeat(setup=A__ , stmt=A__ , repeat=5 , number=1_0000 ) a__ : Optional[int] = repeat(setup=A__ , stmt=A__ , repeat=5 , number=1_0000 ) return (min(A__ ), min(A__ )) if __name__ == "__main__": from doctest import testmod testmod() lowercase : Optional[int] = solution_times() print(F"""The time for naive implementation is {times[0]}.""") print(F"""The time for optimized implementation is {times[1]}.""")
369
import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process lowercase : Union[str, Any] = logging.getLogger(__name__) @dataclass class A__ : """simple docstring""" __A : str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) __A : Optional[str] = field( default=__UpperCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) __A : Optional[str] = field( default='''NER''' , metadata={'''help''': '''Task type to fine tune in training (e.g. NER, POS, etc)'''} ) __A : Optional[str] = field( default=__UpperCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) __A : bool = field(default=__UpperCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __A : Optional[str] = field( default=__UpperCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class A__ : """simple docstring""" __A : str = field( metadata={'''help''': '''The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task.'''} ) __A : Optional[str] = field( default=__UpperCAmelCase , metadata={'''help''': '''Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.'''} , ) __A : int = field( default=1_2_8 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) __A : bool = field( default=__UpperCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def A_ ( ) -> Dict: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. a__ : List[str] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. a__ , a__ , a__ : int = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: a__ , a__ , a__ : List[Any] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. Use' ' --overwrite_output_dir to overcome.' ) a__ : Optional[Any] = import_module('tasks' ) try: a__ : List[Any] = getattr(A__ , model_args.task_type ) a__ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( F'Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ' F'Available tasks classes are: {TokenClassificationTask.__subclasses__()}' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , A__ ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task a__ : Tuple = token_classification_task.get_labels(data_args.labels ) a__ : Dict[int, str] = dict(enumerate(A__ ) ) a__ : Union[str, Any] = len(A__ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. a__ : Optional[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=A__ , idalabel=A__ , labelaid={label: i for i, label in enumerate(A__ )} , cache_dir=model_args.cache_dir , ) a__ : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast , ) a__ : List[Any] = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=A__ , cache_dir=model_args.cache_dir , ) # Get datasets a__ : int = ( TokenClassificationDataset( token_classification_task=A__ , data_dir=data_args.data_dir , tokenizer=A__ , labels=A__ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) a__ : Optional[int] = ( TokenClassificationDataset( token_classification_task=A__ , data_dir=data_args.data_dir , tokenizer=A__ , labels=A__ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def align_predictions(A__ , A__ ) -> Tuple[List[int], List[int]]: a__ : Union[str, Any] = np.argmax(A__ , axis=2 ) a__ , a__ : Dict = preds.shape a__ : Union[str, Any] = [[] for _ in range(A__ )] a__ : Optional[int] = [[] for _ in range(A__ )] for i in range(A__ ): for j in range(A__ ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(A__ ) -> Dict: a__ , a__ : Union[str, Any] = align_predictions(p.predictions , p.label_ids ) return { "accuracy_score": accuracy_score(A__ , A__ ), "precision": precision_score(A__ , A__ ), "recall": recall_score(A__ , A__ ), "f1": fa_score(A__ , A__ ), } # Data collator a__ : Union[str, Any] = DataCollatorWithPadding(A__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer a__ : List[str] = Trainer( model=A__ , args=A__ , train_dataset=A__ , eval_dataset=A__ , compute_metrics=A__ , data_collator=A__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation a__ : Any = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) a__ : Optional[Any] = trainer.evaluate() a__ : List[Any] = os.path.join(training_args.output_dir , 'eval_results.txt' ) if trainer.is_world_process_zero(): with open(A__ , 'w' ) as writer: logger.info('***** Eval results *****' ) for key, value in result.items(): logger.info(' %s = %s' , A__ , A__ ) writer.write('%s = %s\n' % (key, value) ) results.update(A__ ) # Predict if training_args.do_predict: a__ : Optional[Any] = TokenClassificationDataset( token_classification_task=A__ , data_dir=data_args.data_dir , tokenizer=A__ , labels=A__ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.test , ) a__ , a__ , a__ : Any = trainer.predict(A__ ) a__ , a__ : Union[str, Any] = align_predictions(A__ , A__ ) a__ : Optional[int] = os.path.join(training_args.output_dir , 'test_results.txt' ) if trainer.is_world_process_zero(): with open(A__ , 'w' ) as writer: for key, value in metrics.items(): logger.info(' %s = %s' , A__ , A__ ) writer.write('%s = %s\n' % (key, value) ) # Save predictions a__ : Tuple = os.path.join(training_args.output_dir , 'test_predictions.txt' ) if trainer.is_world_process_zero(): with open(A__ , 'w' ) as writer: with open(os.path.join(data_args.data_dir , 'test.txt' ) , 'r' ) as f: token_classification_task.write_predictions_to_file(A__ , A__ , A__ ) return results def A_ ( A__ ) -> Tuple: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
225
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE_: Dict ={ 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: Tuple =[ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: int =[ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys SCREAMING_SNAKE_CASE_: Optional[Any] =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
1
'''simple docstring''' # Function to print upper half of diamond (pyramid) def __snake_case( _lowerCAmelCase ) -> Any: for i in range(0 , _lowerCAmelCase ): for _ in range(0 , n - i - 1 ): # printing spaces print(""" """ , end="""""" ) for _ in range(0 , i + 1 ): # printing stars print("""* """ , end="""""" ) print() def __snake_case( _lowerCAmelCase ) -> List[str]: for i in range(_lowerCAmelCase , 0 , -1 ): for _ in range(_lowerCAmelCase , 0 , -1 ): # printing stars print("""* """ , end="""""" ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(""" """ , end="""""" ) def __snake_case( _lowerCAmelCase ) -> List[Any]: if n <= 0: print(""" ... .... nothing printing :(""" ) return floyd(_lowerCAmelCase ) # upper half reverse_floyd(_lowerCAmelCase ) # lower half if __name__ == "__main__": print(R"| /\ | |- | |- |--| |\ /| |-") print(R"|/ \| |- |_ |_ |__| | \/ | |_") __a = 1 while K: __a = int(input("enter the number and , and see the magic : ")) print() pretty_print(user_number) __a = int(input("press 0 to exit... and 1 to continue...")) print("Good Bye...")
35
0
'''simple docstring''' class __lowercase : '''simple docstring''' def __init__(self ) -> Optional[int]: '''simple docstring''' __lowercase = {} def _UpperCAmelCase (self ) -> None: '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_lowerCamelCase ,''' -> ''' ,''' -> '''.join([str(_lowerCamelCase ) for j in self.vertex[i]] ) ) def _UpperCAmelCase (self ,_lowerCamelCase ,_lowerCamelCase ) -> None: '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_lowerCamelCase ) else: # else make a new vertex __lowercase = [to_vertex] def _UpperCAmelCase (self ) -> None: '''simple docstring''' __lowercase = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_lowerCamelCase ,_lowerCamelCase ) def _UpperCAmelCase (self ,_lowerCamelCase ,_lowerCamelCase ) -> None: '''simple docstring''' __lowercase = True print(_lowerCamelCase ,end=''' ''' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_lowerCamelCase ,_lowerCamelCase ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('''DFS:''') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
217
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _SCREAMING_SNAKE_CASE = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
217
1
"""simple docstring""" from collections import defaultdict class __lowercase : '''simple docstring''' def __init__( self , _UpperCAmelCase , _UpperCAmelCase ): __a : Optional[Any] = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 __a : Any = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(a__ ) ) ] __a : Dict = defaultdict(a__ ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 __a : Union[str, Any] = (1 << len(a__ )) - 1 def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase ): # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement __a : Any = self.count_ways_until(a__ , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. __a : List[str] = total_ways_util return self.dp[mask][task_no] def _lowerCamelCase ( self , _UpperCAmelCase ): # Store the list of persons for each task for i in range(len(a__ ) ): for j in task_performed[i]: self.task[j].append(a__ ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": A = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. A = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
160
"""simple docstring""" from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any] ,_lowerCamelCase : int ) -> List[str]: _lowerCAmelCase : Tuple = k_size // 2 _lowerCAmelCase , _lowerCAmelCase : List[str] = mgrid[0 - center : k_size - center, 0 - center : k_size - center] _lowerCAmelCase : Union[str, Any] = 1 / (2 * pi * sigma) * exp(-(square(_lowerCamelCase ) + square(_lowerCamelCase )) / (2 * square(_lowerCamelCase )) ) return g def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Dict ,_lowerCamelCase : int ,_lowerCamelCase : int ) -> Dict: _lowerCAmelCase , _lowerCAmelCase : str = image.shape[0], image.shape[1] # dst image height and width _lowerCAmelCase : Optional[int] = height - k_size + 1 _lowerCAmelCase : Dict = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows _lowerCAmelCase : Tuple = zeros((dst_height * dst_width, k_size * k_size) ) _lowerCAmelCase : int = 0 for i, j in product(range(_lowerCamelCase ) ,range(_lowerCamelCase ) ): _lowerCAmelCase : Any = ravel(image[i : i + k_size, j : j + k_size] ) _lowerCAmelCase : Union[str, Any] = window row += 1 # turn the kernel into shape(k*k, 1) _lowerCAmelCase : List[Any] = gen_gaussian_kernel(_lowerCamelCase ,_lowerCamelCase ) _lowerCAmelCase : str = ravel(_lowerCamelCase ) # reshape and get the dst image _lowerCAmelCase : int = dot(_lowerCamelCase ,_lowerCamelCase ).reshape(_lowerCamelCase ,_lowerCamelCase ).astype(_lowerCamelCase ) return dst if __name__ == "__main__": # read original image _a : Optional[Any] = imread(r'../image_data/lena.jpg') # turn image in gray scale value _a : Dict = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size _a : Union[str, Any] = gaussian_filter(gray, 3, sigma=1) _a : List[Any] = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow('gaussian filter with 3x3 mask', gaussianaxa) imshow('gaussian filter with 5x5 mask', gaussianaxa) waitKey()
44
0
import qiskit def lowerCamelCase_ ( UpperCamelCase__ : int = 2 ): '''simple docstring''' UpperCamelCase__ = qubits # Using Aer's simulator UpperCamelCase__ = qiskit.Aer.get_backend('''aer_simulator''' ) # Creating a Quantum Circuit acting on the q register UpperCamelCase__ = qiskit.QuantumCircuit(UpperCamelCase__, UpperCamelCase__ ) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0 ) for i in range(1, UpperCamelCase__ ): # Adding CX (CNOT) gate circuit.cx(i - 1, UpperCamelCase__ ) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(UpperCamelCase__ ) ), list(range(UpperCamelCase__ ) ) ) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator UpperCamelCase__ = qiskit.execute(UpperCamelCase__, UpperCamelCase__, shots=1000 ) return job.result().get_counts(UpperCamelCase__ ) if __name__ == "__main__": print(f'Total count for various states are: {quantum_entanglement(3)}')
35
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers lowercase = [ """python""", """tqdm""", """regex""", """requests""", """packaging""", """filelock""", """numpy""", """tokenizers""", """huggingface-hub""", """safetensors""", """accelerate""", """pyyaml""", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py') def lowerCamelCase_ ( UpperCamelCase__ : Union[str, Any], UpperCamelCase__ : Dict=None ): '''simple docstring''' require_version(deps[pkg], UpperCamelCase__ )
35
1
'''simple docstring''' from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
164
'''simple docstring''' from __future__ import annotations import collections import pprint from pathlib import Path def _A ( lowercase__ ): return "".join(sorted(lowercase__ ) ) def _A ( lowercase__ ): return word_by_signature[signature(lowercase__ )] __A = Path(__file__).parent.joinpath("words.txt").read_text(encoding="utf-8") __A = sorted({word.strip().lower() for word in data.splitlines()}) __A = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": __A = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("anagrams.txt", "w") as file: file.write("all_anagrams = \n ") file.write(pprint.pformat(all_anagrams))
164
1
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def __lowerCamelCase ( lowerCAmelCase__ = True , *lowerCAmelCase__ , **lowerCAmelCase__ ): if not is_tqdm_available(): raise ImportError('Accelerate\'s `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.' ) lowerCAmelCase__ = False if main_process_only: lowerCAmelCase__ = PartialState().local_process_index == 0 return _tqdm(*lowerCAmelCase__ , **lowerCAmelCase__ , disable=lowerCAmelCase__ )
119
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = 42 @flax_register_to_config class a_ ( nn.Module , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = 32 UpperCAmelCase_ = 4 UpperCAmelCase_ = 4 UpperCAmelCase_ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) UpperCAmelCase_ = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") UpperCAmelCase_ = False UpperCAmelCase_ = (320, 640, 1_280, 1_280) UpperCAmelCase_ = 2 UpperCAmelCase_ = 8 UpperCAmelCase_ = None UpperCAmelCase_ = 1_280 UpperCAmelCase_ = 0.0 UpperCAmelCase_ = False UpperCAmelCase_ = jnp.floataa UpperCAmelCase_ = True UpperCAmelCase_ = 0 UpperCAmelCase_ = False def __snake_case ( self : Optional[int] , lowercase__ : jax.random.KeyArray): '''simple docstring''' lowerCAmelCase__ = (1, self.in_channels, self.sample_size, self.sample_size) lowerCAmelCase__ = jnp.zeros(lowercase__ , dtype=jnp.floataa) lowerCAmelCase__ = jnp.ones((1,) , dtype=jnp.intaa) lowerCAmelCase__ = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa) lowerCAmelCase__ , lowerCAmelCase__ = jax.random.split(lowercase__) lowerCAmelCase__ = {'params': params_rng, 'dropout': dropout_rng} return self.init(lowercase__ , lowercase__ , lowercase__ , lowercase__)["params"] def __snake_case ( self : Union[str, Any]): '''simple docstring''' lowerCAmelCase__ = self.block_out_channels lowerCAmelCase__ = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( 'At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.') # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. lowerCAmelCase__ = self.num_attention_heads or self.attention_head_dim # input lowerCAmelCase__ = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time lowerCAmelCase__ = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift) lowerCAmelCase__ = FlaxTimestepEmbedding(lowercase__ , dtype=self.dtype) lowerCAmelCase__ = self.only_cross_attention if isinstance(lowercase__ , lowercase__): lowerCAmelCase__ = (only_cross_attention,) * len(self.down_block_types) if isinstance(lowercase__ , lowercase__): lowerCAmelCase__ = (num_attention_heads,) * len(self.down_block_types) # down lowerCAmelCase__ = [] lowerCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types): lowerCAmelCase__ = output_channel lowerCAmelCase__ = block_out_channels[i] lowerCAmelCase__ = i == len(lowercase__) - 1 if down_block_type == "CrossAttnDownBlock2D": lowerCAmelCase__ = FlaxCrossAttnDownBlockaD( in_channels=lowercase__ , out_channels=lowercase__ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: lowerCAmelCase__ = FlaxDownBlockaD( in_channels=lowercase__ , out_channels=lowercase__ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(lowercase__) lowerCAmelCase__ = down_blocks # mid lowerCAmelCase__ = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up lowerCAmelCase__ = [] lowerCAmelCase__ = list(reversed(lowercase__)) lowerCAmelCase__ = list(reversed(lowercase__)) lowerCAmelCase__ = list(reversed(lowercase__)) lowerCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types): lowerCAmelCase__ = output_channel lowerCAmelCase__ = reversed_block_out_channels[i] lowerCAmelCase__ = reversed_block_out_channels[min(i + 1 , len(lowercase__) - 1)] lowerCAmelCase__ = i == len(lowercase__) - 1 if up_block_type == "CrossAttnUpBlock2D": lowerCAmelCase__ = FlaxCrossAttnUpBlockaD( in_channels=lowercase__ , out_channels=lowercase__ , prev_output_channel=lowercase__ , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: lowerCAmelCase__ = FlaxUpBlockaD( in_channels=lowercase__ , out_channels=lowercase__ , prev_output_channel=lowercase__ , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(lowercase__) lowerCAmelCase__ = output_channel lowerCAmelCase__ = up_blocks # out lowerCAmelCase__ = nn.GroupNorm(num_groups=32 , epsilon=1e-5) lowerCAmelCase__ = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Optional[Any] , lowercase__ : Optional[int] , lowercase__ : List[str] , lowercase__ : int , lowercase__ : List[Any]=None , lowercase__ : Union[str, Any]=None , lowercase__ : bool = True , lowercase__ : bool = False , ): '''simple docstring''' if not isinstance(lowercase__ , jnp.ndarray): lowerCAmelCase__ = jnp.array([timesteps] , dtype=jnp.intaa) elif isinstance(lowercase__ , jnp.ndarray) and len(timesteps.shape) == 0: lowerCAmelCase__ = timesteps.astype(dtype=jnp.floataa) lowerCAmelCase__ = jnp.expand_dims(lowercase__ , 0) lowerCAmelCase__ = self.time_proj(lowercase__) lowerCAmelCase__ = self.time_embedding(lowercase__) # 2. pre-process lowerCAmelCase__ = jnp.transpose(lowercase__ , (0, 2, 3, 1)) lowerCAmelCase__ = self.conv_in(lowercase__) # 3. down lowerCAmelCase__ = (sample,) for down_block in self.down_blocks: if isinstance(lowercase__ , lowercase__): lowerCAmelCase__ , lowerCAmelCase__ = down_block(lowercase__ , lowercase__ , lowercase__ , deterministic=not train) else: lowerCAmelCase__ , lowerCAmelCase__ = down_block(lowercase__ , lowercase__ , deterministic=not train) down_block_res_samples += res_samples if down_block_additional_residuals is not None: lowerCAmelCase__ = () for down_block_res_sample, down_block_additional_residual in zip( lowercase__ , lowercase__): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) lowerCAmelCase__ = new_down_block_res_samples # 4. mid lowerCAmelCase__ = self.mid_block(lowercase__ , lowercase__ , lowercase__ , deterministic=not train) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: lowerCAmelCase__ = down_block_res_samples[-(self.layers_per_block + 1) :] lowerCAmelCase__ = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(lowercase__ , lowercase__): lowerCAmelCase__ = up_block( lowercase__ , temb=lowercase__ , encoder_hidden_states=lowercase__ , res_hidden_states_tuple=lowercase__ , deterministic=not train , ) else: lowerCAmelCase__ = up_block(lowercase__ , temb=lowercase__ , res_hidden_states_tuple=lowercase__ , deterministic=not train) # 6. post-process lowerCAmelCase__ = self.conv_norm_out(lowercase__) lowerCAmelCase__ = nn.silu(lowercase__) lowerCAmelCase__ = self.conv_out(lowercase__) lowerCAmelCase__ = jnp.transpose(lowercase__ , (0, 3, 1, 2)) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=lowercase__)
119
1
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np __UpperCAmelCase : str = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 __UpperCAmelCase : int = typing.Union[np.floataa, int, float] # noqa: UP007 def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> VectorOut: return np.sqrt(np.sum((np.asarray(SCREAMING_SNAKE_CASE__) - np.asarray(SCREAMING_SNAKE_CASE__)) ** 2)) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> VectorOut: return sum((va - va) ** 2 for va, va in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__)) ** (1 / 2) if __name__ == "__main__": def A__ ( ) -> None: from timeit import timeit print("""Without Numpy""") print( timeit( """euclidean_distance_no_np([1, 2, 3], [4, 5, 6])""" , number=1_0000 , globals=globals() , )) print("""With Numpy""") print( timeit( """euclidean_distance([1, 2, 3], [4, 5, 6])""" , number=1_0000 , globals=globals() , )) benchmark()
111
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
111
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: __UpperCAmelCase : Any = None __UpperCAmelCase : Dict = logging.get_logger(__name__) __UpperCAmelCase : List[str] = {"vocab_file": "sentencepiece.bpe.model", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase : Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, "tokenizer_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json" ), }, } __UpperCAmelCase : Optional[int] = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } __UpperCAmelCase : Optional[int] = "▁" class UpperCAmelCase_ ( _a): '''simple docstring''' __UpperCamelCase : str = VOCAB_FILES_NAMES __UpperCamelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : str = ["input_ids", "attention_mask"] __UpperCamelCase : int = BarthezTokenizer def __init__( self , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE="<s>" , __SCREAMING_SNAKE_CASE="</s>" , __SCREAMING_SNAKE_CASE="</s>" , __SCREAMING_SNAKE_CASE="<s>" , __SCREAMING_SNAKE_CASE="<unk>" , __SCREAMING_SNAKE_CASE="<pad>" , __SCREAMING_SNAKE_CASE="<mask>" , **__SCREAMING_SNAKE_CASE , ): """simple docstring""" UpperCamelCase : Union[str, Any] = AddedToken(__SCREAMING_SNAKE_CASE , lstrip=__SCREAMING_SNAKE_CASE , rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else mask_token super().__init__( __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) UpperCamelCase : Optional[Any] = vocab_file UpperCamelCase : Tuple = False if not self.vocab_file else True def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] UpperCamelCase : List[str] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase : str = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
315
def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] ): """simple docstring""" UpperCamelCase : list[list[float]] = [] for data in source_data: for i, el in enumerate(SCREAMING_SNAKE_CASE_ ): if len(SCREAMING_SNAKE_CASE_ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(SCREAMING_SNAKE_CASE_ ) ) return data_lists def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] , SCREAMING_SNAKE_CASE_ : list[int] ): """simple docstring""" UpperCamelCase : list[list[float]] = [] for dlist, weight in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : Union[str, Any] = min(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = max(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : list[float] = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: UpperCamelCase : Dict = F"""Invalid weight of {weight:f} provided""" raise ValueError(SCREAMING_SNAKE_CASE_ ) score_lists.append(SCREAMING_SNAKE_CASE_ ) return score_lists def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] ): """simple docstring""" UpperCamelCase : list[float] = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = final_scores[j] + ele return final_scores def a ( SCREAMING_SNAKE_CASE_ : list[list[float]] , SCREAMING_SNAKE_CASE_ : list[int] ): """simple docstring""" UpperCamelCase : str = get_data(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = calculate_each_score(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = generate_final_scores(SCREAMING_SNAKE_CASE_ ) # append scores to source data for i, ele in enumerate(SCREAMING_SNAKE_CASE_ ): source_data[i].append(SCREAMING_SNAKE_CASE_ ) return source_data
315
1
"""simple docstring""" def _snake_case ( _snake_case : str ) -> str: '''simple docstring''' return "".join(chr(ord(_snake_case ) - 32 ) if 'a' <= char <= 'z' else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
315
"""simple docstring""" import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowerCAmelCase ) , '''Tatoeba directory does not exist.''' ) class lowercase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self : Optional[Any] ): _A = tempfile.mkdtemp() return TatoebaConverter(save_dir=_UpperCAmelCase ) @slow def lowerCAmelCase_ ( self : Optional[int] ): self.resolver.convert_models(['heb-eng'] ) @slow def lowerCAmelCase_ ( self : Optional[Any] ): _A , _A = self.resolver.write_model_card('opus-mt-he-en' , dry_run=_UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
315
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( __A : int , __A : int ) -> str: _SCREAMING_SNAKE_CASE = [] create_all_state(1 , A__ , A__ , [] , A__ ) return result def SCREAMING_SNAKE_CASE_ ( __A : int , __A : int , __A : int , __A : list[int] , __A : list[list[int]] , ) -> List[str]: if level == 0: total_list.append(current_list[:] ) return for i in range(A__ , total_number - level + 2 ): current_list.append(A__ ) create_all_state(i + 1 , A__ , level - 1 , A__ , A__ ) current_list.pop() def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] ) -> Any: for i in total_list: print(*A__ ) if __name__ == "__main__": lowerCamelCase_ = 4 lowerCamelCase_ = 2 lowerCamelCase_ = generate_all_combinations(n, k) print_all_state(total_list)
357
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCamelCase_ = { 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['LayoutLMv2FeatureExtractor'] lowerCamelCase_ = ['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
111
0
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) __lowerCamelCase : Dict = logging.getLogger() def __SCREAMING_SNAKE_CASE ( ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() parser.add_argument("""-f""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args() return args.f class __snake_case ( A__ ): def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = logging.StreamHandler(sys.stdout ) logger.addHandler(__snake_case ) def __a ( self : str , _lowercase : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , """run_glue_deebert.py""" ) with patch.object(__snake_case , """argv""" , __snake_case ): SCREAMING_SNAKE_CASE__ = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(__snake_case , 0.6_66 ) @slow @require_torch_non_multi_gpu def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = ''' --model_type roberta --model_name_or_path roberta-base --task_name MRPC --do_train --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --max_seq_length 128 --per_gpu_eval_batch_size=1 --per_gpu_train_batch_size=8 --learning_rate 2e-4 --num_train_epochs 3 --overwrite_output_dir --seed 42 --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --save_steps 0 --overwrite_cache --eval_after_first_stage '''.split() self.run_and_check(__snake_case ) SCREAMING_SNAKE_CASE__ = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --eval_each_highway --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(__snake_case ) SCREAMING_SNAKE_CASE__ = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --early_exit_entropy 0.1 --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(__snake_case )
219
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
0
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def snake_case ( snake_case__ :List[str] , snake_case__ :Dict=False) -> List[str]: try: _A = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _A = default else: # KEY is set, convert it to True or False. try: _A = strtobool(lowerCamelCase__) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''') return _value _SCREAMING_SNAKE_CASE = parse_flag_from_env('RUN_SLOW', default=False) def snake_case ( snake_case__ :Optional[int]) -> Dict: return unittest.skip("""Test was skipped""")(lowerCamelCase__) def snake_case ( snake_case__ :Dict) -> str: return unittest.skipUnless(_run_slow_tests , """test is slow""")(lowerCamelCase__) def snake_case ( snake_case__ :Any) -> Optional[int]: return unittest.skipUnless(not torch.cuda.is_available() , """test requires only a CPU""")(lowerCamelCase__) def snake_case ( snake_case__ :int) -> Optional[int]: return unittest.skipUnless(torch.cuda.is_available() , """test requires a GPU""")(lowerCamelCase__) def snake_case ( snake_case__ :Optional[Any]) -> int: return unittest.skipUnless(is_xpu_available() , """test requires a XPU""")(lowerCamelCase__) def snake_case ( snake_case__ :int) -> Dict: return unittest.skipUnless(is_mps_available() , """test requires a `mps` backend support in `torch`""")(lowerCamelCase__) def snake_case ( snake_case__ :Union[str, Any]) -> str: return unittest.skipUnless( is_transformers_available() and is_datasets_available() , """test requires the Hugging Face suite""")(lowerCamelCase__) def snake_case ( snake_case__ :Dict) -> Optional[int]: return unittest.skipUnless(is_bnb_available() , """test requires the bitsandbytes library""")(lowerCamelCase__) def snake_case ( snake_case__ :Any) -> List[Any]: return unittest.skipUnless(is_tpu_available() , """test requires TPU""")(lowerCamelCase__) def snake_case ( snake_case__ :Optional[Any]) -> Optional[int]: return unittest.skipUnless(torch.cuda.device_count() == 1 , """test requires a GPU""")(lowerCamelCase__) def snake_case ( snake_case__ :int) -> Optional[int]: return unittest.skipUnless(torch.xpu.device_count() == 1 , """test requires a XPU""")(lowerCamelCase__) def snake_case ( snake_case__ :Any) -> str: return unittest.skipUnless(torch.cuda.device_count() > 1 , """test requires multiple GPUs""")(lowerCamelCase__) def snake_case ( snake_case__ :Optional[int]) -> Optional[int]: return unittest.skipUnless(torch.xpu.device_count() > 1 , """test requires multiple XPUs""")(lowerCamelCase__) def snake_case ( snake_case__ :Dict) -> Tuple: return unittest.skipUnless(is_safetensors_available() , """test requires safetensors""")(lowerCamelCase__) def snake_case ( snake_case__ :List[str]) -> List[str]: return unittest.skipUnless(is_deepspeed_available() , """test requires DeepSpeed""")(lowerCamelCase__) def snake_case ( snake_case__ :Union[str, Any]) -> int: return unittest.skipUnless(is_torch_version(""">=""" , """1.12.0""") , """test requires torch version >= 1.12.0""")(lowerCamelCase__) def snake_case ( snake_case__ :List[Any]=None , snake_case__ :int=None) -> Optional[Any]: if test_case is None: return partial(lowerCamelCase__ , version=lowerCamelCase__) return unittest.skipUnless(is_torch_version(""">=""" , lowerCamelCase__) , F'''test requires torch version >= {version}''')(lowerCamelCase__) def snake_case ( snake_case__ :List[str]) -> Optional[int]: return unittest.skipUnless(is_tensorboard_available() , """test requires Tensorboard""")(lowerCamelCase__) def snake_case ( snake_case__ :Optional[Any]) -> Tuple: return unittest.skipUnless(is_wandb_available() , """test requires wandb""")(lowerCamelCase__) def snake_case ( snake_case__ :Optional[int]) -> List[str]: return unittest.skipUnless(is_comet_ml_available() , """test requires comet_ml""")(lowerCamelCase__) _SCREAMING_SNAKE_CASE = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def snake_case ( snake_case__ :List[Any]) -> Union[str, Any]: return unittest.skipUnless( _atleast_one_tracker_available , """test requires at least one tracker to be available and for `comet_ml` to not be installed""" , )(lowerCamelCase__) class a ( unittest.TestCase ): """simple docstring""" lowerCamelCase :Tuple = True @classmethod def UpperCAmelCase ( cls ) -> Optional[int]: _A = tempfile.mkdtemp() @classmethod def UpperCAmelCase ( cls ) -> List[Any]: if os.path.exists(cls.tmpdir ): shutil.rmtree(cls.tmpdir ) def UpperCAmelCase ( self ) -> int: if self.clear_on_setup: for path in Path(self.tmpdir ).glob("""**/*""" ): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(_lowerCamelCase ) class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Optional[int]: _A = mocks if isinstance(_lowerCamelCase , (tuple, list) ) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop ) def snake_case ( snake_case__ :Optional[Any]) -> Dict: _A = AcceleratorState() _A = tensor[None].clone().to(state.device) _A = gather(lowerCamelCase__).cpu() _A = tensor[0].cpu() for i in range(tensors.shape[0]): if not torch.equal(tensors[i] , lowerCamelCase__): return False return True class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _A = returncode _A = stdout _A = stderr async def snake_case ( snake_case__ :Dict , snake_case__ :List[Any]) -> Optional[Any]: while True: _A = await stream.readline() if line: callback(lowerCamelCase__) else: break async def snake_case ( snake_case__ :str , snake_case__ :Union[str, Any]=None , snake_case__ :Optional[int]=None , snake_case__ :Optional[Any]=None , snake_case__ :int=False , snake_case__ :str=False) -> _RunOutput: if echo: print("""\nRunning: """ , """ """.join(lowerCamelCase__)) _A = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCamelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCamelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _A = [] _A = [] def tee(snake_case__ :Optional[int] , snake_case__ :Optional[Any] , snake_case__ :int , snake_case__ :Union[str, Any]=""): _A = line.decode("""utf-8""").rstrip() sink.append(lowerCamelCase__) if not quiet: print(lowerCamelCase__ , lowerCamelCase__ , file=lowerCamelCase__) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda snake_case__: tee(lowerCamelCase__ , lowerCamelCase__ , sys.stdout , label="""stdout:"""))), asyncio.create_task(_read_stream(p.stderr , lambda snake_case__: tee(lowerCamelCase__ , lowerCamelCase__ , sys.stderr , label="""stderr:"""))), ] , timeout=lowerCamelCase__ , ) return _RunOutput(await p.wait() , lowerCamelCase__ , lowerCamelCase__) def snake_case ( snake_case__ :Dict , snake_case__ :List[str]=None , snake_case__ :Optional[int]=None , snake_case__ :str=180 , snake_case__ :Union[str, Any]=False , snake_case__ :List[str]=True) -> _RunOutput: _A = asyncio.get_event_loop() _A = loop.run_until_complete( _stream_subprocess(lowerCamelCase__ , env=lowerCamelCase__ , stdin=lowerCamelCase__ , timeout=lowerCamelCase__ , quiet=lowerCamelCase__ , echo=lowerCamelCase__)) _A = ''' '''.join(lowerCamelCase__) if result.returncode > 0: _A = '''\n'''.join(result.stderr) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''') return result class a ( a__ ): """simple docstring""" pass def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any]=False) -> int: try: _A = subprocess.check_output(lowerCamelCase__ , stderr=subprocess.STDOUT) if return_stdout: if hasattr(lowerCamelCase__ , """decode"""): _A = output.decode("""utf-8""") return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F'''Command `{' '.join(lowerCamelCase__)}` failed with the following error:\n\n{e.output.decode()}''') from e
354
import cva import numpy as np class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: if k in (0.04, 0.06): _A = k _A = window_size else: raise ValueError("""invalid k value""" ) def __str__( self ) -> str: return str(self.k ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> tuple[cva.Mat, list[list[int]]]: _A = cva.imread(lowerCAmelCase_ , 0 ) _A , _A = img.shape _A = [] _A = img.copy() _A = cva.cvtColor(lowerCAmelCase_ , cva.COLOR_GRAY2RGB ) _A , _A = np.gradient(lowerCAmelCase_ ) _A = dx**2 _A = dy**2 _A = dx * dy _A = 0.04 _A = self.window_size // 2 for y in range(lowerCAmelCase_ , h - offset ): for x in range(lowerCAmelCase_ , w - offset ): _A = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = (wxx * wyy) - (wxy**2) _A = wxx + wyy _A = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_55 ) return color_img, corner_list if __name__ == "__main__": _SCREAMING_SNAKE_CASE = HarrisCorner(0.04, 3) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
81
0
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers lowerCAmelCase__ :Optional[int] = [ '''python''', '''tqdm''', '''regex''', '''requests''', '''packaging''', '''filelock''', '''numpy''', '''tokenizers''', '''huggingface-hub''', '''safetensors''', '''accelerate''', '''pyyaml''', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def lowerCAmelCase__ ( a__: Tuple , a__: Optional[int]=None ) -> Any: '''simple docstring''' require_version(deps[pkg] , a__ )
329
import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowerCAmelCase__ ( a__: List[Any] , a__: Union[str, Any]=1_0 ) -> Any: '''simple docstring''' _UpperCAmelCase = [] for _ in range(a__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowerCAmelCase__ ( a__: List[str] , a__: Any=1_0 ) -> List[Any]: '''simple docstring''' _UpperCAmelCase = [] for step in range(a__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase = os.path.join(a__ , 'schedule.bin' ) torch.save(scheduler.state_dict() , a__ ) _UpperCAmelCase = torch.load(a__ ) scheduler.load_state_dict(a__ ) return lrs @require_torch class __a ( unittest.TestCase ): def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) ) for a, b in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertAlmostEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , delta=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self ) -> List[Any]: """simple docstring""" _UpperCAmelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = torch.tensor([0.4, 0.2, -0.5] ) _UpperCAmelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _UpperCAmelCase = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(100 ): _UpperCAmelCase = criterion(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def UpperCAmelCase__ ( self ) -> Dict: """simple docstring""" _UpperCAmelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = torch.tensor([0.4, 0.2, -0.5] ) _UpperCAmelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _UpperCAmelCase = Adafactor( params=[w] , lr=1e-2 , eps=(1e-3_0, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=_SCREAMING_SNAKE_CASE , weight_decay=0.0 , relative_step=_SCREAMING_SNAKE_CASE , scale_parameter=_SCREAMING_SNAKE_CASE , warmup_init=_SCREAMING_SNAKE_CASE , ) for _ in range(1000 ): _UpperCAmelCase = criterion(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class __a ( unittest.TestCase ): _a : Dict = nn.Linear(50 , 50 ) if is_torch_available() else None _a : Dict = AdamW(m.parameters() , lr=1_0.0 ) if is_torch_available() else None _a : List[Any] = 10 def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) -> str: """simple docstring""" self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) ) for a, b in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertAlmostEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , delta=_SCREAMING_SNAKE_CASE , msg=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self ) -> Dict: """simple docstring""" _UpperCAmelCase = {'num_warmup_steps': 2, 'num_training_steps': 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _UpperCAmelCase = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {'num_warmup_steps': 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, 'num_cycles': 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, 'power': 2.0, 'lr_end': 1e-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {'num_warmup_steps': 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): _UpperCAmelCase , _UpperCAmelCase = data _UpperCAmelCase = scheduler_func(self.optimizer , **_SCREAMING_SNAKE_CASE ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _UpperCAmelCase = unwrap_schedule(_SCREAMING_SNAKE_CASE , self.num_steps ) self.assertListAlmostEqual( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , tol=1e-2 , msg=f'''failed for {scheduler_func} in normal scheduler''' , ) _UpperCAmelCase = scheduler_func(self.optimizer , **_SCREAMING_SNAKE_CASE ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(_SCREAMING_SNAKE_CASE ) # wrap to test picklability of the schedule _UpperCAmelCase = unwrap_and_save_reload_schedule(_SCREAMING_SNAKE_CASE , self.num_steps ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , msg=f'''failed for {scheduler_func} in save and reload''' ) class __a : def __init__( self , _SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" _UpperCAmelCase = fn def __call__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" return self.fn(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @classmethod def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" _UpperCAmelCase = list(map(self , scheduler.lr_lambdas ) )
329
1
def __lowerCamelCase ( __a :int ): """simple docstring""" stooge(_lowerCamelCase , 0 , len(_lowerCamelCase ) - 1 ) return arr def __lowerCamelCase ( __a :Dict , __a :Optional[Any] , __a :Optional[int] ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: A__ = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: A__ = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(_lowerCamelCase , _lowerCamelCase , (h - t) ) # Recursively sort last 2/3 elements stooge(_lowerCamelCase , i + t , (_lowerCamelCase) ) # Recursively sort first 2/3 elements stooge(_lowerCamelCase , _lowerCamelCase , (h - t) ) if __name__ == "__main__": A : List[Any] = input('''Enter numbers separated by a comma:\n''').strip() A : int = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
370
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging A : Optional[Any] = logging.get_logger(__name__) A : List[str] = { '''huggingface/time-series-transformer-tourism-monthly''': ( '''https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json''' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Any = '''time_series_transformer''' __lowerCamelCase : Dict = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self : Union[str, Any] , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : str = "student_t" , __lowerCAmelCase : str = "nll" , __lowerCAmelCase : int = 1 , __lowerCAmelCase : List[int] = [1, 2, 3, 4, 5, 6, 7] , __lowerCAmelCase : Optional[Union[str, bool]] = "mean" , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : Optional[List[int]] = None , __lowerCAmelCase : Optional[List[int]] = None , __lowerCAmelCase : int = 32 , __lowerCAmelCase : int = 32 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : bool = True , __lowerCAmelCase : str = "gelu" , __lowerCAmelCase : int = 64 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : int = 1_00 , __lowerCAmelCase : float = 0.0_2 , __lowerCAmelCase : Optional[Any]=True , **__lowerCAmelCase : Union[str, Any] , ) -> Optional[int]: """simple docstring""" A__ = prediction_length A__ = context_length or prediction_length A__ = distribution_output A__ = loss A__ = input_size A__ = num_time_features A__ = lags_sequence A__ = scaling A__ = num_dynamic_real_features A__ = num_static_real_features A__ = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(__lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) A__ = cardinality else: A__ = [0] if embedding_dimension and num_static_categorical_features > 0: if len(__lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) A__ = embedding_dimension else: A__ = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] A__ = num_parallel_samples # Transformer architecture configuration A__ = input_size * len(__lowerCAmelCase ) + self._number_of_features A__ = d_model A__ = encoder_attention_heads A__ = decoder_attention_heads A__ = encoder_ffn_dim A__ = decoder_ffn_dim A__ = encoder_layers A__ = decoder_layers A__ = dropout A__ = attention_dropout A__ = activation_dropout A__ = encoder_layerdrop A__ = decoder_layerdrop A__ = activation_function A__ = init_std A__ = use_cache super().__init__(is_encoder_decoder=__lowerCAmelCase , **__lowerCAmelCase ) @property def a_ ( self : List[Any] ) -> int: """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
276
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class a : def __init__( self : Any , lowerCAmelCase : List[str] , lowerCAmelCase : Optional[Any]=13 , lowerCAmelCase : Any=7 , lowerCAmelCase : Optional[int]=True , lowerCAmelCase : Optional[int]=True , lowerCAmelCase : Optional[int]=True , lowerCAmelCase : Optional[int]=True , lowerCAmelCase : Optional[Any]=99 , lowerCAmelCase : Tuple=[1, 1, 2] , lowerCAmelCase : Any=1 , lowerCAmelCase : int=32 , lowerCAmelCase : Union[str, Any]=4 , lowerCAmelCase : str=8 , lowerCAmelCase : str=37 , lowerCAmelCase : List[str]="gelu_new" , lowerCAmelCase : Union[str, Any]=0.1 , lowerCAmelCase : int=0.1 , lowerCAmelCase : List[Any]=0.0 , lowerCAmelCase : Any=512 , lowerCAmelCase : Optional[int]=3 , lowerCAmelCase : Optional[Any]=0.0_2 , lowerCAmelCase : int=3 , lowerCAmelCase : int=4 , lowerCAmelCase : Dict=None , lowerCAmelCase : int=False , ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE_: List[Any] =parent SCREAMING_SNAKE_CASE_: Union[str, Any] =batch_size SCREAMING_SNAKE_CASE_: Tuple =seq_length SCREAMING_SNAKE_CASE_: List[str] =is_training SCREAMING_SNAKE_CASE_: str =use_input_mask SCREAMING_SNAKE_CASE_: Tuple =use_token_type_ids SCREAMING_SNAKE_CASE_: Dict =use_labels SCREAMING_SNAKE_CASE_: Dict =vocab_size SCREAMING_SNAKE_CASE_: str =block_sizes SCREAMING_SNAKE_CASE_: Tuple =num_decoder_layers SCREAMING_SNAKE_CASE_: Dict =d_model SCREAMING_SNAKE_CASE_: int =n_head SCREAMING_SNAKE_CASE_: Optional[Any] =d_head SCREAMING_SNAKE_CASE_: List[str] =d_inner SCREAMING_SNAKE_CASE_: List[Any] =hidden_act SCREAMING_SNAKE_CASE_: Union[str, Any] =hidden_dropout SCREAMING_SNAKE_CASE_: int =attention_dropout SCREAMING_SNAKE_CASE_: List[str] =activation_dropout SCREAMING_SNAKE_CASE_: Any =max_position_embeddings SCREAMING_SNAKE_CASE_: Any =type_vocab_size SCREAMING_SNAKE_CASE_: int =2 SCREAMING_SNAKE_CASE_: str =num_labels SCREAMING_SNAKE_CASE_: Dict =num_choices SCREAMING_SNAKE_CASE_: Tuple =scope SCREAMING_SNAKE_CASE_: Union[str, Any] =initializer_std # Used in the tests to check the size of the first attention layer SCREAMING_SNAKE_CASE_: int =n_head # Used in the tests to check the size of the first hidden state SCREAMING_SNAKE_CASE_: Optional[int] =self.d_model # Used in the tests to check the number of output hidden states/attentions SCREAMING_SNAKE_CASE_: Tuple =sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: SCREAMING_SNAKE_CASE_: Dict =self.num_hidden_layers + 2 def lowerCamelCase__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE_: Union[str, Any] =None if self.use_input_mask: SCREAMING_SNAKE_CASE_: str =random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE_: Any =None if self.use_token_type_ids: SCREAMING_SNAKE_CASE_: Dict =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE_: str =None SCREAMING_SNAKE_CASE_: Union[str, Any] =None SCREAMING_SNAKE_CASE_: Optional[int] =None if self.use_labels: SCREAMING_SNAKE_CASE_: Optional[Any] =ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_: int =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE_: Optional[Any] =ids_tensor([self.batch_size] , self.num_choices ) SCREAMING_SNAKE_CASE_: List[str] =FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase__ ( self : Optional[int] , lowerCAmelCase : Tuple , lowerCAmelCase : int , lowerCAmelCase : List[Any] , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Dict , lowerCAmelCase : str , lowerCAmelCase : Optional[Any] , ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Dict =TFFunnelModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: List[Any] =model(lowerCAmelCase ) SCREAMING_SNAKE_CASE_: str =[input_ids, input_mask] SCREAMING_SNAKE_CASE_: Union[str, Any] =model(lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) SCREAMING_SNAKE_CASE_: int =False SCREAMING_SNAKE_CASE_: Any =TFFunnelModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Union[str, Any] =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) SCREAMING_SNAKE_CASE_: Any =False SCREAMING_SNAKE_CASE_: Union[str, Any] =TFFunnelModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase__ ( self : Any , lowerCAmelCase : Tuple , lowerCAmelCase : List[str] , lowerCAmelCase : Dict , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : str , lowerCAmelCase : str , lowerCAmelCase : Any , ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =TFFunnelBaseModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Dict ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: List[Any] =model(lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple =[input_ids, input_mask] SCREAMING_SNAKE_CASE_: Dict =model(lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Dict =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) SCREAMING_SNAKE_CASE_: Optional[int] =False SCREAMING_SNAKE_CASE_: int =TFFunnelBaseModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: int =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) SCREAMING_SNAKE_CASE_: Any =False SCREAMING_SNAKE_CASE_: Dict =TFFunnelBaseModel(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Any =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase__ ( self : Dict , lowerCAmelCase : Optional[int] , lowerCAmelCase : Any , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Optional[int] , lowerCAmelCase : Optional[int] , lowerCAmelCase : List[Any] , lowerCAmelCase : Any , ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =TFFunnelForPreTraining(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: List[Any] =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self : str , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : int , lowerCAmelCase : Optional[int] , lowerCAmelCase : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : Any , ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =TFFunnelForMaskedLM(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: str ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: List[str] =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self : List[str] , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : Optional[int] , lowerCAmelCase : Dict , lowerCAmelCase : str , lowerCAmelCase : str , lowerCAmelCase : List[Any] , ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =self.num_labels SCREAMING_SNAKE_CASE_: Any =TFFunnelForSequenceClassification(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: Any =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self : Dict , lowerCAmelCase : Dict , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : int , lowerCAmelCase : List[Any] , lowerCAmelCase : int , lowerCAmelCase : Any , lowerCAmelCase : List[str] , ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE_: str =self.num_choices SCREAMING_SNAKE_CASE_: int =TFFunnelForMultipleChoice(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[Any] =tf.tile(tf.expand_dims(lowerCAmelCase , 1 ) , (1, self.num_choices, 1) ) SCREAMING_SNAKE_CASE_: Dict =tf.tile(tf.expand_dims(lowerCAmelCase , 1 ) , (1, self.num_choices, 1) ) SCREAMING_SNAKE_CASE_: List[str] =tf.tile(tf.expand_dims(lowerCAmelCase , 1 ) , (1, self.num_choices, 1) ) SCREAMING_SNAKE_CASE_: Tuple ={ """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } SCREAMING_SNAKE_CASE_: Optional[Any] =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self : Optional[int] , lowerCAmelCase : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : List[str] , lowerCAmelCase : int , lowerCAmelCase : List[Any] , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Any , ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE_: List[str] =self.num_labels SCREAMING_SNAKE_CASE_: List[Any] =TFFunnelForTokenClassification(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: Optional[Any] =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self : Any , lowerCAmelCase : str , lowerCAmelCase : int , lowerCAmelCase : List[Any] , lowerCAmelCase : Any , lowerCAmelCase : int , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Any , ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =TFFunnelForQuestionAnswering(config=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: int ={"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} SCREAMING_SNAKE_CASE_: Union[str, Any] =model(lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self : List[str] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: str =self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ): Optional[int] =config_and_inputs SCREAMING_SNAKE_CASE_: Tuple ={"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class a ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): UpperCamelCase : Tuple = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) UpperCamelCase : Any = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) UpperCamelCase : str = False UpperCamelCase : List[str] = False def lowerCamelCase__ ( self : List[str] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE_: List[str] =TFFunnelModelTester(self ) SCREAMING_SNAKE_CASE_: Tuple =ConfigTester(self , config_class=lowerCAmelCase ) def lowerCamelCase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self : Union[str, Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def lowerCamelCase__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCAmelCase ) def lowerCamelCase__ ( self : Tuple ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase ) def lowerCamelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE_: Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase ) def lowerCamelCase__ ( self : List[Any] ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE_: Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase ) @require_tf class a ( UpperCAmelCase__ , unittest.TestCase ): UpperCamelCase : List[str] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) UpperCamelCase : Union[str, Any] = False UpperCamelCase : List[Any] = False def lowerCamelCase__ ( self : List[Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =TFFunnelModelTester(self , base=lowerCAmelCase ) SCREAMING_SNAKE_CASE_: str =ConfigTester(self , config_class=lowerCAmelCase ) def lowerCamelCase__ ( self : Dict ) -> str: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self : List[str] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowerCAmelCase ) def lowerCamelCase__ ( self : str ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase ) def lowerCamelCase__ ( self : Optional[int] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_: List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCAmelCase )
173
"""simple docstring""" import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="""session""" ) def __magic_name__ ( ): SCREAMING_SNAKE_CASE_: List[Any] =10 SCREAMING_SNAKE_CASE_: Dict =datasets.Features( { """tokens""": datasets.Sequence(datasets.Value("""string""" ) ), """labels""": datasets.Sequence(datasets.ClassLabel(names=["""negative""", """positive"""] ) ), """answers""": datasets.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), """id""": datasets.Value("""int64""" ), } ) SCREAMING_SNAKE_CASE_: Tuple =datasets.Dataset.from_dict( { """tokens""": [["""foo"""] * 5] * n, """labels""": [[1] * 5] * n, """answers""": [{"""answer_start""": [97], """text""": ["""1976"""]}] * 10, """id""": list(range(lowercase ) ), } , features=lowercase , ) return dataset @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =str(tmp_path_factory.mktemp("""data""" ) / """file.arrow""" ) dataset.map(cache_file_name=lowercase ) return filename # FILE_CONTENT + files _UpperCAmelCase = """\ Text data. Second line of data.""" @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: int =tmp_path_factory.mktemp("""data""" ) / """file.txt""" SCREAMING_SNAKE_CASE_: str =FILE_CONTENT with open(lowercase , """w""" ) as f: f.write(lowercase ) return filename @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): import bza SCREAMING_SNAKE_CASE_: List[str] =tmp_path_factory.mktemp("""data""" ) / """file.txt.bz2""" SCREAMING_SNAKE_CASE_: Union[str, Any] =bytes(lowercase , """utf-8""" ) with bza.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): import gzip SCREAMING_SNAKE_CASE_: List[str] =str(tmp_path_factory.mktemp("""data""" ) / """file.txt.gz""" ) SCREAMING_SNAKE_CASE_: Dict =bytes(lowercase , """utf-8""" ) with gzip.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): if datasets.config.LZ4_AVAILABLE: import lza.frame SCREAMING_SNAKE_CASE_: Tuple =tmp_path_factory.mktemp("""data""" ) / """file.txt.lz4""" SCREAMING_SNAKE_CASE_: List[Any] =bytes(lowercase , """utf-8""" ) with lza.frame.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): if datasets.config.PY7ZR_AVAILABLE: import pyazr SCREAMING_SNAKE_CASE_: Tuple =tmp_path_factory.mktemp("""data""" ) / """file.txt.7z""" with pyazr.SevenZipFile(lowercase , """w""" ) as archive: archive.write(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): import tarfile SCREAMING_SNAKE_CASE_: List[Any] =tmp_path_factory.mktemp("""data""" ) / """file.txt.tar""" with tarfile.TarFile(lowercase , """w""" ) as f: f.add(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): import lzma SCREAMING_SNAKE_CASE_: List[str] =tmp_path_factory.mktemp("""data""" ) / """file.txt.xz""" SCREAMING_SNAKE_CASE_: List[Any] =bytes(lowercase , """utf-8""" ) with lzma.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): import zipfile SCREAMING_SNAKE_CASE_: str =tmp_path_factory.mktemp("""data""" ) / """file.txt.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd SCREAMING_SNAKE_CASE_: Tuple =tmp_path_factory.mktemp("""data""" ) / """file.txt.zst""" SCREAMING_SNAKE_CASE_: Dict =bytes(lowercase , """utf-8""" ) with zstd.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Dict =tmp_path_factory.mktemp("""data""" ) / """file.xml""" SCREAMING_SNAKE_CASE_: Union[str, Any] =textwrap.dedent( """\ <?xml version=\"1.0\" encoding=\"UTF-8\" ?> <tmx version=\"1.4\"> <header segtype=\"sentence\" srclang=\"ca\" /> <body> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv> </tu> </body> </tmx>""" ) with open(lowercase , """w""" ) as f: f.write(lowercase ) return filename _UpperCAmelCase = [ {"""col_1""": """0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """3""", """col_2""": 3, """col_3""": 3.0}, ] _UpperCAmelCase = [ {"""col_1""": """4""", """col_2""": 4, """col_3""": 4.0}, {"""col_1""": """5""", """col_2""": 5, """col_3""": 5.0}, ] _UpperCAmelCase = { """col_1""": ["""0""", """1""", """2""", """3"""], """col_2""": [0, 1, 2, 3], """col_3""": [0.0, 1.0, 2.0, 3.0], } _UpperCAmelCase = [ {"""col_3""": 0.0, """col_1""": """0""", """col_2""": 0}, {"""col_3""": 1.0, """col_1""": """1""", """col_2""": 1}, ] _UpperCAmelCase = [ {"""col_1""": """s0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """s1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """s2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """s3""", """col_2""": 3, """col_3""": 3.0}, ] @pytest.fixture(scope="""session""" ) def __magic_name__ ( ): return DATA_DICT_OF_LISTS @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Tuple =datasets.Dataset.from_dict(lowercase ) SCREAMING_SNAKE_CASE_: Optional[Any] =str(tmp_path_factory.mktemp("""data""" ) / """dataset.arrow""" ) dataset.map(cache_file_name=lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Tuple =str(tmp_path_factory.mktemp("""data""" ) / """dataset.sqlite""" ) with contextlib.closing(sqlitea.connect(lowercase ) ) as con: SCREAMING_SNAKE_CASE_: int =con.cursor() cur.execute("""CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)""" ) for item in DATA: cur.execute("""INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)""" , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Tuple =str(tmp_path_factory.mktemp("""data""" ) / """dataset.csv""" ) with open(lowercase , """w""" , newline="""""" ) as f: SCREAMING_SNAKE_CASE_: int =csv.DictWriter(lowercase , fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[str] =str(tmp_path_factory.mktemp("""data""" ) / """dataset2.csv""" ) with open(lowercase , """w""" , newline="""""" ) as f: SCREAMING_SNAKE_CASE_: Tuple =csv.DictWriter(lowercase , fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): import bza SCREAMING_SNAKE_CASE_: Optional[Any] =tmp_path_factory.mktemp("""data""" ) / """dataset.csv.bz2""" with open(lowercase , """rb""" ) as f: SCREAMING_SNAKE_CASE_: Optional[int] =f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(lowercase , """wb""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: int =tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(lowercase ) ) f.write(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Any =tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(csv_path.replace(""".csv""" , """.CSV""" ) ) ) f.write(lowercase , arcname=os.path.basename(csva_path.replace(""".csv""" , """.CSV""" ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Optional[Any] =tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.csv.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: str =str(tmp_path_factory.mktemp("""data""" ) / """dataset.parquet""" ) SCREAMING_SNAKE_CASE_: List[Any] =pa.schema( { """col_1""": pa.string(), """col_2""": pa.intaa(), """col_3""": pa.floataa(), } ) with open(lowercase , """wb""" ) as f: SCREAMING_SNAKE_CASE_: int =pq.ParquetWriter(lowercase , schema=lowercase ) SCREAMING_SNAKE_CASE_: str =pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(lowercase ) )] for k in DATA[0]} , schema=lowercase ) writer.write_table(lowercase ) writer.close() return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Optional[int] =str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) SCREAMING_SNAKE_CASE_: Optional[Any] ={"""data""": DATA} with open(lowercase , """w""" ) as f: json.dump(lowercase , lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) SCREAMING_SNAKE_CASE_: Tuple ={"""data""": DATA_DICT_OF_LISTS} with open(lowercase , """w""" ) as f: json.dump(lowercase , lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Any =str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl""" ) with open(lowercase , """w""" ) as f: for item in DATA: f.write(json.dumps(lowercase ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Optional[Any] =str(tmp_path_factory.mktemp("""data""" ) / """dataset2.jsonl""" ) with open(lowercase , """w""" ) as f: for item in DATA: f.write(json.dumps(lowercase ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Optional[int] =str(tmp_path_factory.mktemp("""data""" ) / """dataset_312.jsonl""" ) with open(lowercase , """w""" ) as f: for item in DATA_312: f.write(json.dumps(lowercase ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Tuple =str(tmp_path_factory.mktemp("""data""" ) / """dataset-str.jsonl""" ) with open(lowercase , """w""" ) as f: for item in DATA_STR: f.write(json.dumps(lowercase ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): import gzip SCREAMING_SNAKE_CASE_: Any =str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt.gz""" ) with open(lowercase , """rb""" ) as orig_file: with gzip.open(lowercase , """wb""" ) as zipped_file: zipped_file.writelines(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): import gzip SCREAMING_SNAKE_CASE_: int =str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.gz""" ) with open(lowercase , """rb""" ) as orig_file: with gzip.open(lowercase , """wb""" ) as zipped_file: zipped_file.writelines(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(lowercase ) ) f.write(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Dict =tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.join("""nested""" , os.path.basename(lowercase ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Optional[int] =tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.jsonl.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Dict =tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.tar""" with tarfile.TarFile(lowercase , """w""" ) as f: f.add(lowercase , arcname=os.path.basename(lowercase ) ) f.add(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: int =tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.tar""" with tarfile.TarFile(lowercase , """w""" ) as f: f.add(lowercase , arcname=os.path.join("""nested""" , os.path.basename(lowercase ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =["""0""", """1""", """2""", """3"""] SCREAMING_SNAKE_CASE_: Optional[Any] =str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt""" ) with open(lowercase , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Dict =["""0""", """1""", """2""", """3"""] SCREAMING_SNAKE_CASE_: Dict =str(tmp_path_factory.mktemp("""data""" ) / """dataset2.txt""" ) with open(lowercase , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Optional[Any] =["""0""", """1""", """2""", """3"""] SCREAMING_SNAKE_CASE_: List[str] =tmp_path_factory.mktemp("""data""" ) / """dataset.abc""" with open(lowercase , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Union[str, Any] =tmp_path_factory.mktemp("""data""" ) / """dataset.text.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(lowercase ) ) f.write(lowercase , arcname=os.path.basename(lowercase ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: int =tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.text.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) f.write(lowercase , arcname=os.path.join("""main_dir""" , os.path.basename(lowercase ) ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Union[str, Any] =tmp_path_factory.mktemp("""data""" ) / """dataset.ext.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename("""unsupported.ext""" ) ) f.write(lowercase , arcname=os.path.basename("""unsupported_2.ext""" ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[str] ="""\n""".join(["""First""", """Second\u2029with Unicode new line""", """Third"""] ) SCREAMING_SNAKE_CASE_: List[Any] =str(tmp_path_factory.mktemp("""data""" ) / """dataset_with_unicode_new_lines.txt""" ) with open(lowercase , """w""" , encoding="""utf-8""" ) as f: f.write(lowercase ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( ): return os.path.join("""tests""" , """features""" , """data""" , """test_image_rgb.jpg""" ) @pytest.fixture(scope="""session""" ) def __magic_name__ ( ): return os.path.join("""tests""" , """features""" , """data""" , """test_audio_44100.wav""" ) @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Any =tmp_path_factory.mktemp("""data""" ) / """dataset.img.zip""" with zipfile.ZipFile(lowercase , """w""" ) as f: f.write(lowercase , arcname=os.path.basename(lowercase ) ) f.write(lowercase , arcname=os.path.basename(lowercase ).replace(""".jpg""" , """2.jpg""" ) ) return path @pytest.fixture(scope="""session""" ) def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[str] =tmp_path_factory.mktemp("""data_dir""" ) (data_dir / "subdir").mkdir() with open(data_dir / """subdir""" / """train.txt""" , """w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """subdir""" / """test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) # hidden file with open(data_dir / """subdir""" / """.test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / """.subdir""" / """train.txt""" , """w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """.subdir""" / """test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) return data_dir
173
1
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowerCAmelCase__ : Optional[Any] =argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--original_config_file', type=str, required=True, help='The YAML config file corresponding to the original architecture.', ) parser.add_argument( '--num_in_channels', default=None, type=int, help='The number of input channels. If `None` number of input channels will be automatically inferred.', ) parser.add_argument( '--image_size', default=5_12, type=int, help=( 'The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2' ' Base. Use 768 for Stable Diffusion v2.' ), ) parser.add_argument( '--extract_ema', action='store_true', help=( 'Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights' ' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield' ' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.' ), ) parser.add_argument( '--upcast_attention', action='store_true', help=( 'Whether the attention computation should always be upcasted. This is necessary when running stable' ' diffusion 2.1.' ), ) parser.add_argument( '--from_safetensors', action='store_true', help='If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.', ) parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') def a__ ( A__ ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( '--use_linear_projection', help='Override for use linear projection', required=False, type=parse_bool ) parser.add_argument('--cross_attention_dim', help='Override for cross attention_dim', required=False, type=int) lowerCAmelCase__ : Dict =parser.parse_args() lowerCAmelCase__ : Any =download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
352
lowerCAmelCase__ : Dict ='ABCDEFGHIJKLMNOPQRSTUVWXYZ' def a__ ( ): SCREAMING_SNAKE_CASE_ : Tuple = input('Enter message: ' ) SCREAMING_SNAKE_CASE_ : Any = input('Enter key [alphanumeric]: ' ) SCREAMING_SNAKE_CASE_ : Tuple = input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'encrypt' SCREAMING_SNAKE_CASE_ : int = encrypt_message(A__, A__ ) elif mode.lower().startswith('d' ): SCREAMING_SNAKE_CASE_ : List[Any] = 'decrypt' SCREAMING_SNAKE_CASE_ : Any = decrypt_message(A__, A__ ) print(F'''\n{mode.title()}ed message:''' ) print(A__ ) def a__ ( A__, A__ ): return translate_message(A__, A__, 'encrypt' ) def a__ ( A__, A__ ): return translate_message(A__, A__, 'decrypt' ) def a__ ( A__, A__, A__ ): SCREAMING_SNAKE_CASE_ : Tuple = [] SCREAMING_SNAKE_CASE_ : str = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = key.upper() for symbol in message: SCREAMING_SNAKE_CASE_ : List[str] = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(A__ ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(A__ ): SCREAMING_SNAKE_CASE_ : Dict = 0 else: translated.append(A__ ) return "".join(A__ ) if __name__ == "__main__": main()
162
0
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class A__ ( unittest.TestCase , __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = load_tool("text-classification" ) self.tool.setup() UpperCamelCase : List[str] = load_tool("text-classification" , remote=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.tool("That's quite cool" , ["positive", "negative"] ) self.assertEqual(A_ , "positive" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.remote_tool("That's quite cool" , ["positive", "negative"] ) self.assertEqual(A_ , "positive" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.tool(text="That's quite cool" , labels=["positive", "negative"] ) self.assertEqual(A_ , "positive" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.remote_tool(text="That's quite cool" , labels=["positive", "negative"] ) self.assertEqual(A_ , "positive" )
52
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=24 , A_=2 , A_=6 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=None , A_=1000 , ): '''simple docstring''' UpperCamelCase : Union[str, Any] = parent UpperCamelCase : List[Any] = batch_size UpperCamelCase : Dict = seq_length UpperCamelCase : Tuple = is_training UpperCamelCase : Union[str, Any] = use_input_mask UpperCamelCase : Tuple = use_token_type_ids UpperCamelCase : Optional[Any] = use_labels UpperCamelCase : str = vocab_size UpperCamelCase : Optional[int] = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Optional[Any] = num_attention_heads UpperCamelCase : Optional[Any] = intermediate_size UpperCamelCase : Optional[Any] = hidden_act UpperCamelCase : Union[str, Any] = hidden_dropout_prob UpperCamelCase : Union[str, Any] = attention_probs_dropout_prob UpperCamelCase : List[Any] = max_position_embeddings UpperCamelCase : str = type_vocab_size UpperCamelCase : Optional[int] = type_sequence_label_size UpperCamelCase : Dict = initializer_range UpperCamelCase : int = num_labels UpperCamelCase : Optional[int] = scope UpperCamelCase : int = range_bbox def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : Any = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCamelCase : Union[str, Any] = bbox[i, j, 3] UpperCamelCase : int = bbox[i, j, 1] UpperCamelCase : int = t if bbox[i, j, 2] < bbox[i, j, 0]: UpperCamelCase : List[str] = bbox[i, j, 2] UpperCamelCase : Optional[int] = bbox[i, j, 0] UpperCamelCase : Optional[Any] = t UpperCamelCase : Dict = None if self.use_input_mask: UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) UpperCamelCase : str = None if self.use_token_type_ids: UpperCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Dict = None UpperCamelCase : int = None if self.use_labels: UpperCamelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : List[Any] = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def __UpperCamelCase( self ): '''simple docstring''' return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Any = LiltModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : str = model(A_ , bbox=A_ , attention_mask=A_ , token_type_ids=A_ ) UpperCamelCase : Optional[int] = model(A_ , bbox=A_ , token_type_ids=A_ ) UpperCamelCase : Any = model(A_ , bbox=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Any = self.num_labels UpperCamelCase : Dict = LiltForTokenClassification(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Dict = model( A_ , bbox=A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Dict = LiltForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : List[str] = model( A_ , bbox=A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Tuple = config_and_inputs UpperCamelCase : Tuple = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Union[str, Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': LiltModel, 'question-answering': LiltForQuestionAnswering, 'text-classification': LiltForSequenceClassification, 'token-classification': LiltForTokenClassification, 'zero-shot': LiltForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase :Dict = False _UpperCAmelCase :Union[str, Any] = False def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' return True def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = LiltModelTester(self ) UpperCamelCase : Optional[int] = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCamelCase : Union[str, Any] = type self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Dict = LiltModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch @slow class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = LiltModel.from_pretrained("SCUT-DLVCLab/lilt-roberta-en-base" ).to(A_ ) UpperCamelCase : Tuple = torch.tensor([[1, 2]] , device=A_ ) UpperCamelCase : List[str] = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Optional[int] = model(input_ids=A_ , bbox=A_ ) UpperCamelCase : List[str] = torch.Size([1, 2, 768] ) UpperCamelCase : Any = torch.tensor( [[-0.06_53, 0.09_50, -0.00_61], [-0.05_45, 0.09_26, -0.03_24]] , device=A_ , ) self.assertTrue(outputs.last_hidden_state.shape , A_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , A_ , atol=1e-3 ) )
52
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _A = { "configuration_longformer": [ "LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "LongformerConfig", "LongformerOnnxConfig", ], "tokenization_longformer": ["LongformerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["LongformerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "LongformerForMaskedLM", "LongformerForMultipleChoice", "LongformerForQuestionAnswering", "LongformerForSequenceClassification", "LongformerForTokenClassification", "LongformerModel", "LongformerPreTrainedModel", "LongformerSelfAttention", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLongformerForMaskedLM", "TFLongformerForMultipleChoice", "TFLongformerForQuestionAnswering", "TFLongformerForSequenceClassification", "TFLongformerForTokenClassification", "TFLongformerModel", "TFLongformerPreTrainedModel", "TFLongformerSelfAttention", ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys _A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
359
from ....utils import logging _A = logging.get_logger(__name__) class lowerCamelCase ( A_ ): def __init__(self : Tuple , _A : Optional[int] , _A : Tuple=None , _A : Union[str, Any]=2_0_4_8 ) -> List[Any]: snake_case = config.__dict__ snake_case = modal_hidden_size if num_labels: snake_case = num_labels
137
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
311
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase : str = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[str] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
36
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
36
1
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy __lowerCAmelCase : int = logging.getLogger(__name__) __lowerCAmelCase : List[Any] = 'pytorch_model.bin' @dataclasses.dataclass class UpperCAmelCase_ : '''simple docstring''' a__ = dataclasses.field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models."""} ) a__ = dataclasses.field( default=_A , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co."""} , ) @dataclasses.dataclass class UpperCAmelCase_ : '''simple docstring''' a__ = dataclasses.field(metadata={"""help""": """A csv or a json file containing the training data."""} ) a__ = dataclasses.field(metadata={"""help""": """A csv or a json file containing the data to predict on."""} ) a__ = dataclasses.field( default=_A , metadata={"""help""": """A csv or a json file containing the validation data."""} ) a__ = dataclasses.field( default=_A , metadata={"""help""": """The name of the task to train on."""} , ) a__ = dataclasses.field( default=_A , metadata={"""help""": """The list of labels for the task."""} ) @dataclasses.dataclass class UpperCAmelCase_ : '''simple docstring''' a__ = dataclasses.field( metadata={"""help""": """The output directory where the model predictions and checkpoints will be written."""} ) a__ = dataclasses.field( default="""accuracy""" , metadata={"""help""": """The evaluation metric used for the task."""} ) a__ = dataclasses.field( default="""no""" , metadata={ """help""": """The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]""" } , ) a__ = dataclasses.field( default=10 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) a__ = dataclasses.field( default=0.0 , metadata={ """help""": """How much the specified evaluation metric must improve to satisfy early stopping conditions.""" } , ) a__ = dataclasses.field( default=_A , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the confidence score."""} , ) a__ = dataclasses.field( default=_A , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the validation performance."""} , ) a__ = dataclasses.field( default=_A , metadata={"""help""": """Whether to fine-tune on labeled data after pseudo training."""} , ) a__ = dataclasses.field( default=0.0 , metadata={"""help""": """Confidence threshold for pseudo-labeled data filtering."""} , ) a__ = dataclasses.field( default=1_00 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) a__ = dataclasses.field( default=_A , metadata={"""help""": """Random seed for initialization."""} , ) def a__ ( A_, A_, A_, A_, A_, A_ ): '''simple docstring''' __magic_name__ = datasets.concatenate_datasets([infer_input, infer_output], axis=1 ) if args.do_filter_by_confidence: __magic_name__ = dataset.filter(lambda A_ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 __magic_name__ = int(eval_result * len(A_ ) ) print(A_ ) __magic_name__ = dataset.sort("""probability""", reverse=A_ ) __magic_name__ = dataset.select(range(A_ ) ) __magic_name__ = dataset.remove_columns(["""label""", """probability"""] ) __magic_name__ = dataset.rename_column("""prediction""", """label""" ) __magic_name__ = dataset.map(lambda A_ : {"label": idalabel[example["label"]]} ) __magic_name__ = dataset.shuffle(seed=args.seed ) __magic_name__ = os.path.join(A_, f'''train_pseudo.{args.data_file_extension}''' ) if args.data_file_extension == "csv": dataset.to_csv(A_, index=A_ ) else: dataset.to_json(A_ ) def a__ ( A_, A_, A_, A_, **A_ ): '''simple docstring''' __magic_name__ = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO, ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() __magic_name__ = STModelArguments(model_name_or_path=A_ ) __magic_name__ = STDataArguments(train_file=A_, infer_file=A_ ) __magic_name__ = STTrainingArguments(output_dir=A_ ) __magic_name__ = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(A_ ).items(): setattr(A_, A_, A_ ) for key, value in kwargs.items(): if hasattr(A_, A_ ): setattr(A_, A_, A_ ) # Sanity checks __magic_name__ = {} __magic_name__ = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None __magic_name__ = args.train_file __magic_name__ = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None __magic_name__ = args.eval_file for key in data_files: __magic_name__ = data_files[key].split(""".""" )[-1] assert extension in ["csv", "json"], f'''`{key}_file` should be a csv or a json file.''' if args.data_file_extension is None: __magic_name__ = extension else: assert extension == args.data_file_extension, f'''`{key}_file` should be a {args.data_file_extension} file`.''' assert ( args.eval_metric in datasets.list_metrics() ), f'''{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.''' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("""Creating the initial data directory for self-training...""" ) __magic_name__ = f'''{args.output_dir}/self-train_iter-{{}}'''.format __magic_name__ = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir, exist_ok=A_ ) os.makedirs(A_, exist_ok=A_ ) accelerator.wait_for_everyone() __magic_name__ = None __magic_name__ = None __magic_name__ = 0 __magic_name__ = False # Show the progress bar __magic_name__ = tqdm(range(args.max_selftrain_iterations ), disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0, int(args.max_selftrain_iterations ) ): __magic_name__ = data_dir_format(A_ ) assert os.path.exists(A_ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 __magic_name__ = os.path.join(A_, """stage-1""" ) __magic_name__ = { """accelerator""": accelerator, """model_name_or_path""": args.model_name_or_path, """cache_dir""": args.cache_dir, """do_train""": True, """train_file""": data_files["""train"""] if iteration == 0 else data_files["""train_pseudo"""], """do_eval""": True if args.eval_file is not None else False, """eval_file""": data_files["""eval"""], """do_predict""": True, """infer_file""": data_files["""infer"""], """task_name""": args.task_name, """label_list""": args.label_list, """output_dir""": current_output_dir, """eval_metric""": args.eval_metric, """evaluation_strategy""": args.evaluation_strategy, """early_stopping_patience""": args.early_stopping_patience, """early_stopping_threshold""": args.early_stopping_threshold, """seed""": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(A_, A_ ): arguments_dict.update({key: value} ) __magic_name__ = os.path.join(A_, """best-checkpoint""", A_ ) if os.path.exists(A_ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.""", A_, A_, ) else: logger.info("""***** Running self-training: iteration: %d, stage: 1 *****""", A_ ) finetune(**A_ ) accelerator.wait_for_everyone() assert os.path.exists(A_ ) logger.info("""Self-training job completed: iteration: %d, stage: 1.""", A_ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data __magic_name__ = os.path.join(A_, """best-checkpoint""" ) __magic_name__ = os.path.join(A_, """stage-2""" ) # Update arguments_dict __magic_name__ = model_path __magic_name__ = data_files["""train"""] __magic_name__ = current_output_dir __magic_name__ = os.path.join(A_, """best-checkpoint""", A_ ) if os.path.exists(A_ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.""", A_, A_, ) else: logger.info("""***** Running self-training: iteration: %d, stage: 2 *****""", A_ ) finetune(**A_ ) accelerator.wait_for_everyone() assert os.path.exists(A_ ) logger.info("""Self-training job completed: iteration: %d, stage: 2.""", A_ ) __magic_name__ = iteration __magic_name__ = data_dir_format(iteration + 1 ) __magic_name__ = AutoConfig.from_pretrained(os.path.join(A_, """best-checkpoint""" ) ) __magic_name__ = config.idalabel __magic_name__ = os.path.join(A_, """eval_results_best-checkpoint.json""" ) __magic_name__ = os.path.join(A_, """test_results_best-checkpoint.json""" ) assert os.path.exists(A_ ) with open(A_, """r""" ) as f: __magic_name__ = float(json.load(A_ )[args.eval_metric] ) __magic_name__ = os.path.join(A_, """infer_output_best-checkpoint.csv""" ) assert os.path.exists(A_ ) # Loading the dataset from local csv or json files. __magic_name__ = load_dataset(args.data_file_extension, data_files={"""data""": data_files["""infer"""]} )["""data"""] __magic_name__ = load_dataset("""csv""", data_files={"""data""": infer_output_file} )["""data"""] if accelerator.is_main_process: os.makedirs(A_, exist_ok=A_ ) shutil.copy(A_, os.path.join(A_, f'''eval_results_iter-{iteration}.json''' ) ) if os.path.exists(A_ ): shutil.copy(A_, os.path.join(A_, f'''test_results_iter-{iteration}.json''' ) ) create_pseudo_labeled_data(A_, A_, A_, A_, A_, A_ ) accelerator.wait_for_everyone() __magic_name__ = os.path.join(A_, f'''train_pseudo.{args.data_file_extension}''' ) if args.evaluation_strategy != IntervalStrategy.NO.value: __magic_name__ = eval_result if best_iteration is None: __magic_name__ = new_iteration __magic_name__ = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: __magic_name__ = new_iteration __magic_name__ = new_eval_result __magic_name__ = 0 else: if new_eval_result == best_eval_result: __magic_name__ = new_iteration __magic_name__ = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: __magic_name__ = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("""Best iteration: %d""", A_ ) logger.info("""Best evaluation result: %s = %f""", args.eval_metric, A_ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A_, f'''eval_results_iter-{iteration}.json''' ), os.path.join(A_, """eval_results_best-iteration.json""" ), ) else: # Assume that the last iteration is the best logger.info("""Best iteration: %d""", args.max_selftrain_iterations - 1 ) logger.info("""Best evaluation result: %s = %f""", args.eval_metric, A_ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A_, f'''eval_results_iter-{args.max_selftrain_iterations - 1}.json''' ), os.path.join(A_, """eval_results_best-iteration.json""" ), )
88
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ : Dict = { "configuration_blip_2": [ "BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Blip2Config", "Blip2QFormerConfig", "Blip2VisionConfig", ], "processing_blip_2": ["Blip2Processor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = [ "BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST", "Blip2Model", "Blip2QFormerModel", "Blip2PreTrainedModel", "Blip2ForConditionalGeneration", "Blip2VisionModel", ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys A_ : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
165
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class __SCREAMING_SNAKE_CASE ( unittest.TestCase): def __init__( self , _UpperCamelCase , _UpperCamelCase=7 , _UpperCamelCase=3 , _UpperCamelCase=30 , _UpperCamelCase=4_00 , _UpperCamelCase=True , _UpperCamelCase=None , _UpperCamelCase=True , _UpperCamelCase=[0.5, 0.5, 0.5] , _UpperCamelCase=[0.5, 0.5, 0.5] , _UpperCamelCase=True , _UpperCamelCase=1 / 2_55 , _UpperCamelCase=True , ): """simple docstring""" lowerCAmelCase__ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 13_33} lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = min_resolution lowerCAmelCase__ = max_resolution lowerCAmelCase__ = do_resize lowerCAmelCase__ = size lowerCAmelCase__ = do_normalize lowerCAmelCase__ = image_mean lowerCAmelCase__ = image_std lowerCAmelCase__ = do_rescale lowerCAmelCase__ = rescale_factor lowerCAmelCase__ = do_pad def UpperCamelCase__ ( self ): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase=False ): """simple docstring""" if not batched: lowerCAmelCase__ = image_inputs[0] if isinstance(_UpperCamelCase , Image.Image ): lowerCAmelCase__ , lowerCAmelCase__ = image.size else: lowerCAmelCase__ , lowerCAmelCase__ = image.shape[1], image.shape[2] if w < h: lowerCAmelCase__ = int(self.size['shortest_edge'] * h / w ) lowerCAmelCase__ = self.size['shortest_edge'] elif w > h: lowerCAmelCase__ = self.size['shortest_edge'] lowerCAmelCase__ = int(self.size['shortest_edge'] * w / h ) else: lowerCAmelCase__ = self.size['shortest_edge'] lowerCAmelCase__ = self.size['shortest_edge'] else: lowerCAmelCase__ = [] for image in image_inputs: lowerCAmelCase__ , lowerCAmelCase__ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCAmelCase__ = max(_UpperCamelCase , key=lambda _UpperCamelCase : item[0] )[0] lowerCAmelCase__ = max(_UpperCamelCase , key=lambda _UpperCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Any = YolosImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = YolosImageProcessingTester(self ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCamelCase , 'image_mean' ) ) self.assertTrue(hasattr(_UpperCamelCase , 'image_std' ) ) self.assertTrue(hasattr(_UpperCamelCase , 'do_normalize' ) ) self.assertTrue(hasattr(_UpperCamelCase , 'do_resize' ) ) self.assertTrue(hasattr(_UpperCamelCase , 'size' ) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 13_33} ) self.assertEqual(image_processor.do_pad , _UpperCamelCase ) lowerCAmelCase__ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_UpperCamelCase ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , Image.Image ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase , batched=_UpperCamelCase ) lowerCAmelCase__ = image_processing(_UpperCamelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase , numpify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , np.ndarray ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ = image_processing(_UpperCamelCase , return_tensors='pt' ).pixel_values lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase , batched=_UpperCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase , torchify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , torch.Tensor ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ = image_processing(_UpperCamelCase , return_tensors='pt' ).pixel_values lowerCAmelCase__ , lowerCAmelCase__ = self.image_processor_tester.get_expected_values(_UpperCamelCase , batched=_UpperCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) lowerCAmelCase__ = self.image_processing_class(do_resize=_UpperCamelCase , do_normalize=_UpperCamelCase , do_rescale=_UpperCamelCase ) # create random PyTorch tensors lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCamelCase , torchify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowerCAmelCase__ = image_processing_a.pad(_UpperCamelCase , return_tensors='pt' ) lowerCAmelCase__ = image_processing_a(_UpperCamelCase , return_tensors='pt' ) self.assertTrue( torch.allclose(encoded_images_with_method['pixel_values'] , encoded_images['pixel_values'] , atol=1E-4 ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCAmelCase__ = json.loads(f.read() ) lowerCAmelCase__ = {'image_id': 3_97_69, 'annotations': target} # encode them lowerCAmelCase__ = YolosImageProcessor.from_pretrained('hustvl/yolos-small' ) lowerCAmelCase__ = image_processing(images=_UpperCamelCase , annotations=_UpperCamelCase , return_tensors='pt' ) # verify pixel values lowerCAmelCase__ = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['pixel_values'].shape , _UpperCamelCase ) lowerCAmelCase__ = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _UpperCamelCase , atol=1E-4 ) ) # verify area lowerCAmelCase__ = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _UpperCamelCase ) ) # verify boxes lowerCAmelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _UpperCamelCase ) lowerCAmelCase__ = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _UpperCamelCase , atol=1E-3 ) ) # verify image_id lowerCAmelCase__ = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _UpperCamelCase ) ) # verify is_crowd lowerCAmelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _UpperCamelCase ) ) # verify class_labels lowerCAmelCase__ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _UpperCamelCase ) ) # verify orig_size lowerCAmelCase__ = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _UpperCamelCase ) ) # verify size lowerCAmelCase__ = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _UpperCamelCase ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCAmelCase__ = json.loads(f.read() ) lowerCAmelCase__ = {'file_name': '000000039769.png', 'image_id': 3_97_69, 'segments_info': target} lowerCAmelCase__ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCAmelCase__ = YolosImageProcessor(format='coco_panoptic' ) lowerCAmelCase__ = image_processing(images=_UpperCamelCase , annotations=_UpperCamelCase , masks_path=_UpperCamelCase , return_tensors='pt' ) # verify pixel values lowerCAmelCase__ = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['pixel_values'].shape , _UpperCamelCase ) lowerCAmelCase__ = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _UpperCamelCase , atol=1E-4 ) ) # verify area lowerCAmelCase__ = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _UpperCamelCase ) ) # verify boxes lowerCAmelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _UpperCamelCase ) lowerCAmelCase__ = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _UpperCamelCase , atol=1E-3 ) ) # verify image_id lowerCAmelCase__ = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _UpperCamelCase ) ) # verify is_crowd lowerCAmelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _UpperCamelCase ) ) # verify class_labels lowerCAmelCase__ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _UpperCamelCase ) ) # verify masks lowerCAmelCase__ = 82_28_73 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _UpperCamelCase ) # verify orig_size lowerCAmelCase__ = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _UpperCamelCase ) ) # verify size lowerCAmelCase__ = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _UpperCamelCase ) )
370
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __SCREAMING_SNAKE_CASE ( __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Tuple = BarthezTokenizer _SCREAMING_SNAKE_CASE : int = BarthezTokenizerFast _SCREAMING_SNAKE_CASE : Dict = True _SCREAMING_SNAKE_CASE : Tuple = True def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase__ = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_UpperCamelCase ) lowerCAmelCase__ = tokenizer def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = '<pad>' lowerCAmelCase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCamelCase ) , _UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCamelCase ) , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_UpperCamelCase ) , 10_11_22 ) def UpperCamelCase__ ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] lowerCAmelCase__ = [0, 57, 30_18, 7_03_07, 91, 2] lowerCAmelCase__ = self.tokenizer( _UpperCamelCase , max_length=len(_UpperCamelCase ) , padding=_UpperCamelCase , truncation=_UpperCamelCase , return_tensors='pt' ) self.assertIsInstance(_UpperCamelCase , _UpperCamelCase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowerCAmelCase__ = batch.input_ids.tolist()[0] self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" if not self.test_rust_tokenizer: return lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = self.get_rust_tokenizer() lowerCAmelCase__ = 'I was born in 92000, and this is falsé.' lowerCAmelCase__ = tokenizer.tokenize(_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase__ = tokenizer.encode(_UpperCamelCase , add_special_tokens=_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.encode(_UpperCamelCase , add_special_tokens=_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase__ = self.get_rust_tokenizer() lowerCAmelCase__ = tokenizer.encode(_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.encode(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) @slow def UpperCamelCase__ ( self ): """simple docstring""" # fmt: off lowerCAmelCase__ = {'input_ids': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowerCAmelCase__ = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_UpperCamelCase , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=_UpperCamelCase , )
122
0
"""simple docstring""" import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case = logging.get_logger(__name__) __snake_case = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } __snake_case = { '''b0''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def A_ ( _lowerCAmelCase : int ): """simple docstring""" _a = EfficientNetConfig() _a = CONFIG_MAP[model_name]['''hidden_dim'''] _a = CONFIG_MAP[model_name]['''width_coef'''] _a = CONFIG_MAP[model_name]['''depth_coef'''] _a = CONFIG_MAP[model_name]['''image_size'''] _a = CONFIG_MAP[model_name]['''dropout_rate'''] _a = CONFIG_MAP[model_name]['''dw_padding'''] _a = '''huggingface/label-files''' _a = '''imagenet-1k-id2label.json''' _a = 10_00 _a = json.load(open(hf_hub_download(_lowerCAmelCase, _lowerCAmelCase, repo_type='''dataset''' ), '''r''' ) ) _a = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} _a = idalabel _a = {v: k for k, v in idalabel.items()} return config def A_ ( ): """simple docstring""" _a = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _a = Image.open(requests.get(_lowerCAmelCase, stream=_lowerCAmelCase ).raw ) return im def A_ ( _lowerCAmelCase : str ): """simple docstring""" _a = CONFIG_MAP[model_name]['''image_size'''] _a = EfficientNetImageProcessor( size={'''height''': size, '''width''': size}, image_mean=[0.4_8_5, 0.4_5_6, 0.4_0_6], image_std=[0.4_7_8_5_3_9_4_4, 0.4_7_3_2_8_6_4, 0.4_7_4_3_4_1_6_3], do_center_crop=_lowerCAmelCase, ) return preprocessor def A_ ( _lowerCAmelCase : int ): """simple docstring""" _a = [v.split('''_''' )[0].split('''block''' )[1] for v in original_param_names if v.startswith('''block''' )] _a = sorted(set(_lowerCAmelCase ) ) _a = len(_lowerCAmelCase ) _a = {b: str(_lowerCAmelCase ) for b, i in zip(_lowerCAmelCase, range(_lowerCAmelCase ) )} _a = [] rename_keys.append(('''stem_conv/kernel:0''', '''embeddings.convolution.weight''') ) rename_keys.append(('''stem_bn/gamma:0''', '''embeddings.batchnorm.weight''') ) rename_keys.append(('''stem_bn/beta:0''', '''embeddings.batchnorm.bias''') ) rename_keys.append(('''stem_bn/moving_mean:0''', '''embeddings.batchnorm.running_mean''') ) rename_keys.append(('''stem_bn/moving_variance:0''', '''embeddings.batchnorm.running_var''') ) for b in block_names: _a = block_name_mapping[b] rename_keys.append((f'block{b}_expand_conv/kernel:0', f'encoder.blocks.{hf_b}.expansion.expand_conv.weight') ) rename_keys.append((f'block{b}_expand_bn/gamma:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.weight') ) rename_keys.append((f'block{b}_expand_bn/beta:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.bias') ) rename_keys.append( (f'block{b}_expand_bn/moving_mean:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_mean') ) rename_keys.append( (f'block{b}_expand_bn/moving_variance:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_var') ) rename_keys.append( (f'block{b}_dwconv/depthwise_kernel:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight') ) rename_keys.append((f'block{b}_bn/gamma:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight') ) rename_keys.append((f'block{b}_bn/beta:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias') ) rename_keys.append( (f'block{b}_bn/moving_mean:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean') ) rename_keys.append( (f'block{b}_bn/moving_variance:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var') ) rename_keys.append((f'block{b}_se_reduce/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.weight') ) rename_keys.append((f'block{b}_se_reduce/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.bias') ) rename_keys.append((f'block{b}_se_expand/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.weight') ) rename_keys.append((f'block{b}_se_expand/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.bias') ) rename_keys.append( (f'block{b}_project_conv/kernel:0', f'encoder.blocks.{hf_b}.projection.project_conv.weight') ) rename_keys.append((f'block{b}_project_bn/gamma:0', f'encoder.blocks.{hf_b}.projection.project_bn.weight') ) rename_keys.append((f'block{b}_project_bn/beta:0', f'encoder.blocks.{hf_b}.projection.project_bn.bias') ) rename_keys.append( (f'block{b}_project_bn/moving_mean:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_mean') ) rename_keys.append( (f'block{b}_project_bn/moving_variance:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_var') ) rename_keys.append(('''top_conv/kernel:0''', '''encoder.top_conv.weight''') ) rename_keys.append(('''top_bn/gamma:0''', '''encoder.top_bn.weight''') ) rename_keys.append(('''top_bn/beta:0''', '''encoder.top_bn.bias''') ) rename_keys.append(('''top_bn/moving_mean:0''', '''encoder.top_bn.running_mean''') ) rename_keys.append(('''top_bn/moving_variance:0''', '''encoder.top_bn.running_var''') ) _a = {} for item in rename_keys: if item[0] in original_param_names: _a = '''efficientnet.''' + item[1] _a = '''classifier.weight''' _a = '''classifier.bias''' return key_mapping def A_ ( _lowerCAmelCase : str, _lowerCAmelCase : Any, _lowerCAmelCase : List[Any] ): """simple docstring""" for key, value in tf_params.items(): if "normalization" in key: continue _a = key_mapping[key] if "_conv" in key and "kernel" in key: _a = torch.from_numpy(_lowerCAmelCase ).permute(3, 2, 0, 1 ) elif "depthwise_kernel" in key: _a = torch.from_numpy(_lowerCAmelCase ).permute(2, 3, 0, 1 ) elif "kernel" in key: _a = torch.from_numpy(np.transpose(_lowerCAmelCase ) ) else: _a = torch.from_numpy(_lowerCAmelCase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(_lowerCAmelCase ) @torch.no_grad() def A_ ( _lowerCAmelCase : Any, _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Dict, _lowerCAmelCase : str ): """simple docstring""" _a = model_classes[model_name]( include_top=_lowerCAmelCase, weights='''imagenet''', input_tensor=_lowerCAmelCase, input_shape=_lowerCAmelCase, pooling=_lowerCAmelCase, classes=10_00, classifier_activation='''softmax''', ) _a = original_model.trainable_variables _a = original_model.non_trainable_variables _a = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: _a = param.numpy() _a = list(tf_params.keys() ) # Load HuggingFace model _a = get_efficientnet_config(_lowerCAmelCase ) _a = EfficientNetForImageClassification(_lowerCAmelCase ).eval() _a = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print('''Converting parameters...''' ) _a = rename_keys(_lowerCAmelCase ) replace_params(_lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase ) # Initialize preprocessor and preprocess input image _a = convert_image_processor(_lowerCAmelCase ) _a = preprocessor(images=prepare_img(), return_tensors='''pt''' ) # HF model inference hf_model.eval() with torch.no_grad(): _a = hf_model(**_lowerCAmelCase ) _a = outputs.logits.detach().numpy() # Original model inference _a = False _a = CONFIG_MAP[model_name]['''image_size'''] _a = prepare_img().resize((image_size, image_size), resample=PIL.Image.NEAREST ) _a = image.img_to_array(_lowerCAmelCase ) _a = np.expand_dims(_lowerCAmelCase, axis=0 ) _a = original_model.predict(_lowerCAmelCase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(_lowerCAmelCase, _lowerCAmelCase, atol=1e-3 ), "The predicted logits are not the same." print('''Model outputs match!''' ) if save_model: # Create folder to save model if not os.path.isdir(_lowerCAmelCase ): os.mkdir(_lowerCAmelCase ) # Save converted model and image processor hf_model.save_pretrained(_lowerCAmelCase ) preprocessor.save_pretrained(_lowerCAmelCase ) if push_to_hub: # Push model and image processor to hub print(f'Pushing converted {model_name} to the hub...' ) _a = f'efficientnet-{model_name}' preprocessor.push_to_hub(_lowerCAmelCase ) hf_model.push_to_hub(_lowerCAmelCase ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') __snake_case = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
320
"""simple docstring""" def A_ ( ): """simple docstring""" _a = [] _a = 1 while len(_lowerCAmelCase ) < 1e6: constant.append(str(_lowerCAmelCase ) ) i += 1 _a = ''''''.join(_lowerCAmelCase ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[9_99] ) * int(constant[99_99] ) * int(constant[9_99_99] ) * int(constant[99_99_99] ) ) if __name__ == "__main__": print(solution())
320
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = { """configuration_deberta""": ["""DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DebertaConfig""", """DebertaOnnxConfig"""], """tokenization_deberta""": ["""DebertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = ["""DebertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Optional[Any] = [ """DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """DebertaForMaskedLM""", """DebertaForQuestionAnswering""", """DebertaForSequenceClassification""", """DebertaForTokenClassification""", """DebertaModel""", """DebertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = [ """TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDebertaForMaskedLM""", """TFDebertaForQuestionAnswering""", """TFDebertaForSequenceClassification""", """TFDebertaForTokenClassification""", """TFDebertaModel""", """TFDebertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys lowerCamelCase_ : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
350
import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _UpperCamelCase ( _A , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : Tuple = CTRLTokenizer __UpperCamelCase : int = False __UpperCamelCase : List[str] = False def lowerCAmelCase__ ( self : List[str] ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase_: int = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCamelCase_: int = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) UpperCamelCase_: Union[str, Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCamelCase_: Tuple = {"""unk_token""": """<unk>"""} UpperCamelCase_: Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCamelCase_: Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case_ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(snake_case_ ) ) def lowerCAmelCase__ ( self : Optional[int] , **snake_case_ : int ): kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def lowerCAmelCase__ ( self : Optional[int] , snake_case_ : List[str] ): UpperCamelCase_: Dict = """adapt react readapt apt""" UpperCamelCase_: List[str] = """adapt react readapt apt""" return input_text, output_text def lowerCAmelCase__ ( self : List[str] ): UpperCamelCase_: str = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase_: List[Any] = """adapt react readapt apt""" UpperCamelCase_: Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCamelCase_: int = tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) UpperCamelCase_: List[Any] = tokens + [tokenizer.unk_token] UpperCamelCase_: Union[str, Any] = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case_ ) , snake_case_ )
223
0
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase__ = False class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def lowercase__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Tuple = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) lowerCAmelCase : Tuple = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) lowerCAmelCase : Tuple = torch.manual_seed(0 ) lowerCAmelCase : int = pipe.dual_guided( prompt="first prompt" , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(snake_case__ ) lowerCAmelCase : str = VersatileDiffusionPipeline.from_pretrained(snake_case__ , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) lowerCAmelCase : List[Any] = generator.manual_seed(0 ) lowerCAmelCase : Union[str, Any] = pipe.dual_guided( prompt="first prompt" , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) lowerCAmelCase : str = "cyberpunk 2077" lowerCAmelCase : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) lowerCAmelCase : List[Any] = pipe.dual_guided( prompt=snake_case__ , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images lowerCAmelCase : Tuple = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase : Optional[Any] = np.array([0.1448, 0.1619, 0.1741, 0.1086, 0.1147, 0.1128, 0.1199, 0.1165, 0.1001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 lowerCAmelCase : Union[str, Any] = "A painting of a squirrel eating a burger " lowerCAmelCase : Any = torch.manual_seed(0 ) lowerCAmelCase : Any = pipe.text_to_image( prompt=snake_case__ , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images lowerCAmelCase : Dict = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase : Dict = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 lowerCAmelCase : Union[str, Any] = pipe.image_variation(snake_case__ , generator=snake_case__ , output_type="numpy" ).images lowerCAmelCase : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase : Any = np.array([0.3076, 0.3123, 0.3284, 0.3782, 0.3770, 0.3894, 0.4297, 0.4331, 0.4456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
108
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : int ="dpr" def __init__( self , snake_case__=30_522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3_072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=0 , snake_case__="absolute" , snake_case__ = 0 , **snake_case__ , ): """simple docstring""" super().__init__(pad_token_id=snake_case__ , **snake_case__ ) lowerCAmelCase : Union[str, Any] = vocab_size lowerCAmelCase : str = hidden_size lowerCAmelCase : Any = num_hidden_layers lowerCAmelCase : Optional[int] = num_attention_heads lowerCAmelCase : Union[str, Any] = hidden_act lowerCAmelCase : Dict = intermediate_size lowerCAmelCase : Union[str, Any] = hidden_dropout_prob lowerCAmelCase : Dict = attention_probs_dropout_prob lowerCAmelCase : Dict = max_position_embeddings lowerCAmelCase : Tuple = type_vocab_size lowerCAmelCase : Any = initializer_range lowerCAmelCase : Any = layer_norm_eps lowerCAmelCase : Dict = projection_dim lowerCAmelCase : Dict = position_embedding_type
108
1
def lowerCAmelCase_ (lowerCAmelCase__: int = 6_0_0_8_5_1_4_7_5_1_4_3 ): """simple docstring""" try: UpperCAmelCase_: Any = int(lowerCAmelCase__ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) UpperCAmelCase_: Dict = 2 UpperCAmelCase_: List[str] = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 UpperCAmelCase_: Optional[int] = i while n % i == 0: UpperCAmelCase_: Dict = n // i i += 1 return int(lowerCAmelCase__ ) if __name__ == "__main__": print(F'''{solution() = }''')
82
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _a ( _lowerCAmelCase ): def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: Union[str, Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, """embed_dim""" ) ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, """num_heads""" ) ) class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=13, SCREAMING_SNAKE_CASE_=64, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=[16, 48, 96], SCREAMING_SNAKE_CASE_=[1, 3, 6], SCREAMING_SNAKE_CASE_=[1, 2, 10], SCREAMING_SNAKE_CASE_=[7, 3, 3], SCREAMING_SNAKE_CASE_=[4, 2, 2], SCREAMING_SNAKE_CASE_=[2, 1, 1], SCREAMING_SNAKE_CASE_=[2, 2, 2], SCREAMING_SNAKE_CASE_=[False, False, True], SCREAMING_SNAKE_CASE_=[0.0, 0.0, 0.0], SCREAMING_SNAKE_CASE_=0.0_2, SCREAMING_SNAKE_CASE_=1E-12, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=2, ) -> List[Any]: UpperCAmelCase_: Union[str, Any] = parent UpperCAmelCase_: Any = batch_size UpperCAmelCase_: Optional[int] = image_size UpperCAmelCase_: Tuple = patch_sizes UpperCAmelCase_: int = patch_stride UpperCAmelCase_: int = patch_padding UpperCAmelCase_: List[str] = is_training UpperCAmelCase_: List[Any] = use_labels UpperCAmelCase_: int = num_labels UpperCAmelCase_: Dict = num_channels UpperCAmelCase_: Any = embed_dim UpperCAmelCase_: Optional[Any] = num_heads UpperCAmelCase_: Dict = stride_kv UpperCAmelCase_: Dict = depth UpperCAmelCase_: Optional[Any] = cls_token UpperCAmelCase_: List[str] = attention_drop_rate UpperCAmelCase_: List[str] = initializer_range UpperCAmelCase_: Tuple = layer_norm_eps def __snake_case (self ) -> Dict: UpperCAmelCase_: str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_: Optional[Any] = None if self.use_labels: # create a random int32 tensor of given shape UpperCAmelCase_: str = ids_tensor([self.batch_size], self.num_labels ) UpperCAmelCase_: List[str] = self.get_config() return config, pixel_values, labels def __snake_case (self ) -> Tuple: return CvtConfig( image_size=self.image_size, num_labels=self.num_labels, num_channels=self.num_channels, embed_dim=self.embed_dim, num_heads=self.num_heads, patch_sizes=self.patch_sizes, patch_padding=self.patch_padding, patch_stride=self.patch_stride, stride_kv=self.stride_kv, depth=self.depth, cls_token=self.cls_token, attention_drop_rate=self.attention_drop_rate, initializer_range=self.initializer_range, ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCAmelCase_: Optional[int] = TFCvtModel(config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Tuple = model(SCREAMING_SNAKE_CASE_, training=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[Any] = (self.image_size, self.image_size) UpperCAmelCase_ , UpperCAmelCase_: Any = image_size[0], image_size[1] for i in range(len(self.depth ) ): UpperCAmelCase_: Optional[Any] = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) UpperCAmelCase_: str = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.embed_dim[-1], height, width) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> str: UpperCAmelCase_: List[str] = self.num_labels UpperCAmelCase_: Tuple = TFCvtForImageClassification(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Any = model(SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_, training=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def __snake_case (self ) -> Dict: UpperCAmelCase_: Any = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: Union[str, Any] = config_and_inputs UpperCAmelCase_: Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _a ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): A = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () A = ( {'''feature-extraction''': TFCvtModel, '''image-classification''': TFCvtForImageClassification} if is_tf_available() else {} ) A = False A = False A = False A = False A = False def __snake_case (self ) -> int: UpperCAmelCase_: Tuple = TFCvtModelTester(self ) UpperCAmelCase_: Dict = TFCvtConfigTester(self, config_class=SCREAMING_SNAKE_CASE_, has_text_modality=SCREAMING_SNAKE_CASE_, hidden_size=37 ) def __snake_case (self ) -> List[Any]: self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="""Cvt does not output attentions""" ) def __snake_case (self ) -> Optional[int]: pass @unittest.skip(reason="""Cvt does not use inputs_embeds""" ) def __snake_case (self ) -> List[str]: pass @unittest.skip(reason="""Cvt does not support input and output embeddings""" ) def __snake_case (self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0, reason="""TF does not support backprop for grouped convolutions on CPU.""", ) def __snake_case (self ) -> Optional[int]: super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0, reason="""TF does not support backprop for grouped convolutions on CPU.""", ) @slow def __snake_case (self ) -> int: super().test_keras_fit() @unittest.skip(reason="""Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8""" ) def __snake_case (self ) -> List[Any]: UpperCAmelCase_: List[str] = tf.keras.mixed_precision.Policy("""mixed_float16""" ) tf.keras.mixed_precision.set_global_policy(SCREAMING_SNAKE_CASE_ ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("""float32""" ) def __snake_case (self ) -> Tuple: UpperCAmelCase_ , UpperCAmelCase_: str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_: List[str] = model_class(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_: Any = [*signature.parameters.keys()] UpperCAmelCase_: Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1], SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Any: def check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Optional[int] = model_class(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ) UpperCAmelCase_: Optional[Any] = outputs.hidden_states UpperCAmelCase_: Optional[int] = len(self.model_tester.depth ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ), SCREAMING_SNAKE_CASE_ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ), [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ], ) UpperCAmelCase_ , UpperCAmelCase_: Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_: int = True check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_: Tuple = True check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> int: UpperCAmelCase_: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE_ ) @slow def __snake_case (self ) -> Optional[int]: for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_: Union[str, Any] = TFCvtModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _a ( unittest.TestCase ): @cached_property def __snake_case (self ) -> Tuple: return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __snake_case (self ) -> Dict: UpperCAmelCase_: Tuple = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase_: Dict = self.default_image_processor UpperCAmelCase_: Dict = prepare_img() UpperCAmelCase_: Union[str, Any] = image_processor(images=SCREAMING_SNAKE_CASE_, return_tensors="""tf""" ) # forward pass UpperCAmelCase_: int = model(**SCREAMING_SNAKE_CASE_ ) # verify the logits UpperCAmelCase_: Optional[int] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Dict = tf.constant([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy(), SCREAMING_SNAKE_CASE_, atol=1E-4 ) )
82
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available lowerCAmelCase__ = {'configuration_speech_encoder_decoder': ['SpeechEncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['SpeechEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['FlaxSpeechEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
11
"""simple docstring""" import os import pytest from attr import dataclass _a : Optional[int]= "us-east-1" # defaults region @dataclass class UpperCamelCase : UpperCAmelCase : str UpperCAmelCase : Any = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" UpperCAmelCase : Optional[Any] = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 500, """save_steps""": 5500, } UpperCAmelCase : List[Any] = {**hyperparameters, """max_steps""": 1000} @property def _lowercase (self : Dict) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def _lowercase (self : Tuple) -> str: return f"{self.framework}-transfromers-test" @property def _lowercase (self : Any) -> str: return f"./tests/sagemaker/scripts/{self.framework}" @property def _lowercase (self : List[str]) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def __UpperCAmelCase ( UpperCAmelCase_ : Optional[Any] ) -> int: '''simple docstring''' __snake_case : List[Any] = SageMakerTestEnvironment(framework=request.cls.framework )
172
0
import csv import tweepy # Twitter API credentials A__ : Dict ="" A__ : Optional[int] ="" A__ : str ="" A__ : List[str] ="" def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = tweepy.OAuthHandler(lowerCAmelCase , lowerCAmelCase ) auth.set_access_token(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = tweepy.API(lowerCAmelCase ) # initialize a list to hold all the tweepy Tweets _lowerCAmelCase = [] # make initial request for most recent tweets (200 is the maximum allowed count) _lowerCAmelCase = api.user_timeline(screen_name=lowerCAmelCase , count=2_00 ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # save the id of the oldest tweet less one _lowerCAmelCase = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(lowerCAmelCase ) > 0: print(f"getting tweets before {oldest}" ) # all subsequent requests use the max_id param to prevent duplicates _lowerCAmelCase = api.user_timeline( screen_name=lowerCAmelCase , count=2_00 , max_id=lowerCAmelCase ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # update the id of the oldest tweet less one _lowerCAmelCase = alltweets[-1].id - 1 print(f"...{len(lowerCAmelCase )} tweets downloaded so far" ) # transform the tweepy tweets into a 2D array that will populate the csv _lowerCAmelCase = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(f"new_{screen_name}_tweets.csv" , """w""" ) as f: _lowerCAmelCase = csv.writer(lowerCAmelCase ) writer.writerow(["""id""", """created_at""", """text"""] ) writer.writerows(lowerCAmelCase ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
355
'''simple docstring''' # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] _lowerCAmelCase = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } _lowerCAmelCase = f"{src_lang}-{tgt_lang}" _lowerCAmelCase = f"\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"facebook/wmt19-{src_lang}-{tgt_lang}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR's WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n" os.makedirs(lowerCAmelCase , exist_ok=lowerCAmelCase ) _lowerCAmelCase = os.path.join(lowerCAmelCase , """README.md""" ) print(f"Generating {path}" ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(lowerCAmelCase ) # make sure we are under the root of the project A__ : Optional[int] =Path(__file__).resolve().parent.parent.parent A__ : Union[str, Any] =repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: A__ , A__ , A__ : Optional[Any] =model_name.split('''-''') A__ : List[str] =model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
220
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = { "configuration_electra": ["ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP", "ElectraConfig", "ElectraOnnxConfig"], "tokenization_electra": ["ElectraTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["ElectraTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "ElectraForCausalLM", "ElectraForMaskedLM", "ElectraForMultipleChoice", "ElectraForPreTraining", "ElectraForQuestionAnswering", "ElectraForSequenceClassification", "ElectraForTokenClassification", "ElectraModel", "ElectraPreTrainedModel", "load_tf_weights_in_electra", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFElectraForMaskedLM", "TFElectraForMultipleChoice", "TFElectraForPreTraining", "TFElectraForQuestionAnswering", "TFElectraForSequenceClassification", "TFElectraForTokenClassification", "TFElectraModel", "TFElectraPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "FlaxElectraForCausalLM", "FlaxElectraForMaskedLM", "FlaxElectraForMultipleChoice", "FlaxElectraForPreTraining", "FlaxElectraForQuestionAnswering", "FlaxElectraForSequenceClassification", "FlaxElectraForTokenClassification", "FlaxElectraModel", "FlaxElectraPreTrainedModel", ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
35
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class A__ ( lowerCAmelCase__ ): lowerCAmelCase__ : Union[str, Any] = ["vqvae"] def __init__( self : int , _UpperCAmelCase : AutoencoderKL , _UpperCAmelCase : UNetaDConditionModel , _UpperCAmelCase : Mel , _UpperCAmelCase : Union[DDIMScheduler, DDPMScheduler] , ) -> str: """simple docstring""" super().__init__() self.register_modules(unet=_UpperCAmelCase , scheduler=_UpperCAmelCase , mel=_UpperCAmelCase , vqvae=_UpperCAmelCase ) def a__ ( self : Tuple ) -> int: """simple docstring""" return 50 if isinstance(self.scheduler , _UpperCAmelCase ) else 10_00 @torch.no_grad() def __call__( self : str , _UpperCAmelCase : int = 1 , _UpperCAmelCase : str = None , _UpperCAmelCase : np.ndarray = None , _UpperCAmelCase : int = 0 , _UpperCAmelCase : int = 0 , _UpperCAmelCase : int = None , _UpperCAmelCase : torch.Generator = None , _UpperCAmelCase : float = 0 , _UpperCAmelCase : float = 0 , _UpperCAmelCase : torch.Generator = None , _UpperCAmelCase : float = 0 , _UpperCAmelCase : torch.Tensor = None , _UpperCAmelCase : torch.Tensor = None , _UpperCAmelCase : str=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: """simple docstring""" __lowercase = steps or self.get_default_steps() self.scheduler.set_timesteps(_UpperCAmelCase ) __lowercase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: __lowercase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: __lowercase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=_UpperCAmelCase , device=self.device , ) __lowercase = noise __lowercase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(_UpperCAmelCase , _UpperCAmelCase ) __lowercase = self.mel.audio_slice_to_image(_UpperCAmelCase ) __lowercase = np.frombuffer(input_image.tobytes() , dtype='uint8' ).reshape( (input_image.height, input_image.width) ) __lowercase = (input_image / 2_55) * 2 - 1 __lowercase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: __lowercase = self.vqvae.encode(torch.unsqueeze(_UpperCAmelCase , 0 ) ).latent_dist.sample( generator=_UpperCAmelCase )[0] __lowercase = self.vqvae.config.scaling_factor * input_images if start_step > 0: __lowercase = self.scheduler.add_noise(_UpperCAmelCase , _UpperCAmelCase , self.scheduler.timesteps[start_step - 1] ) __lowercase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) __lowercase = int(mask_start_secs * pixels_per_second ) __lowercase = int(mask_end_secs * pixels_per_second ) __lowercase = self.scheduler.add_noise(_UpperCAmelCase , _UpperCAmelCase , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , _UpperCAmelCase ): __lowercase = self.unet(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase )['sample'] else: __lowercase = self.unet(_UpperCAmelCase , _UpperCAmelCase )['sample'] if isinstance(self.scheduler , _UpperCAmelCase ): __lowercase = self.scheduler.step( model_output=_UpperCAmelCase , timestep=_UpperCAmelCase , sample=_UpperCAmelCase , eta=_UpperCAmelCase , generator=_UpperCAmelCase , )['prev_sample'] else: __lowercase = self.scheduler.step( model_output=_UpperCAmelCase , timestep=_UpperCAmelCase , sample=_UpperCAmelCase , generator=_UpperCAmelCase , )['prev_sample'] if mask is not None: if mask_start > 0: __lowercase = mask[:, step, :, :mask_start] if mask_end > 0: __lowercase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance __lowercase = 1 / self.vqvae.config.scaling_factor * images __lowercase = self.vqvae.decode(_UpperCAmelCase )['sample'] __lowercase = (images / 2 + 0.5).clamp(0 , 1 ) __lowercase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() __lowercase = (images * 2_55).round().astype('uint8' ) __lowercase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(_UpperCAmelCase , mode='RGB' ).convert('L' ) for _ in images) ) __lowercase = [self.mel.image_to_audio(_UpperCAmelCase ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(_UpperCAmelCase )[:, np.newaxis, :] ) , **ImagePipelineOutput(_UpperCAmelCase ) ) @torch.no_grad() def a__ ( self : Any , _UpperCAmelCase : List[Image.Image] , _UpperCAmelCase : int = 50 ) -> np.ndarray: """simple docstring""" assert isinstance(self.scheduler , _UpperCAmelCase ) self.scheduler.set_timesteps(_UpperCAmelCase ) __lowercase = np.array( [np.frombuffer(image.tobytes() , dtype='uint8' ).reshape((1, image.height, image.width) ) for image in images] ) __lowercase = (sample / 2_55) * 2 - 1 __lowercase = torch.Tensor(_UpperCAmelCase ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): __lowercase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps __lowercase = self.scheduler.alphas_cumprod[t] __lowercase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) __lowercase = 1 - alpha_prod_t __lowercase = self.unet(_UpperCAmelCase , _UpperCAmelCase )['sample'] __lowercase = (1 - alpha_prod_t_prev) ** 0.5 * model_output __lowercase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) __lowercase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def a__ ( _UpperCAmelCase : torch.Tensor , _UpperCAmelCase : torch.Tensor , _UpperCAmelCase : float ) -> torch.Tensor: """simple docstring""" __lowercase = acos(torch.dot(torch.flatten(_UpperCAmelCase ) , torch.flatten(_UpperCAmelCase ) ) / torch.norm(_UpperCAmelCase ) / torch.norm(_UpperCAmelCase ) ) return sin((1 - alpha) * theta ) * xa / sin(_UpperCAmelCase ) + sin(alpha * theta ) * xa / sin(_UpperCAmelCase )
325
0
"""simple docstring""" import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() _a : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : str ) -> str: print("""Loading config file...""" ) def flatten_yaml_as_dict(_lowerCamelCase : List[Any] ,_lowerCamelCase : Tuple="" ,_lowerCamelCase : List[Any]="." ): _lowerCAmelCase : Union[str, Any] = [] for k, v in d.items(): _lowerCAmelCase : List[str] = parent_key + sep + k if parent_key else k if isinstance(_lowerCamelCase ,collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(_lowerCamelCase ,_lowerCamelCase ,sep=_lowerCamelCase ).items() ) else: items.append((new_key, v) ) return dict(_lowerCamelCase ) _lowerCAmelCase : Optional[Any] = argparse.Namespace() with open(_lowerCamelCase ,"""r""" ) as yaml_file: try: _lowerCAmelCase : int = yaml.load(_lowerCamelCase ,Loader=yaml.FullLoader ) _lowerCAmelCase : Optional[int] = flatten_yaml_as_dict(_lowerCamelCase ) for k, v in flat_cfg.items(): setattr(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) except yaml.YAMLError as exc: logger.error("""Error while loading config file: {}. Error message: {}""".format(_lowerCamelCase ,str(_lowerCamelCase ) ) ) return config def SCREAMING_SNAKE_CASE ( _lowerCamelCase : List[str] ,_lowerCamelCase : Union[str, Any] ) -> Union[str, Any]: _lowerCAmelCase : int = MobileViTVaConfig() _lowerCAmelCase : Optional[Any] = False # dataset if task_name.startswith("""imagenet1k_""" ): _lowerCAmelCase : Union[str, Any] = 1000 if int(task_name.strip().split("""_""" )[-1] ) == 384: _lowerCAmelCase : Any = 384 else: _lowerCAmelCase : Dict = 256 _lowerCAmelCase : Optional[int] = """imagenet-1k-id2label.json""" elif task_name.startswith("""imagenet21k_to_1k_""" ): _lowerCAmelCase : Tuple = 21000 if int(task_name.strip().split("""_""" )[-1] ) == 384: _lowerCAmelCase : List[str] = 384 else: _lowerCAmelCase : str = 256 _lowerCAmelCase : Any = """imagenet-22k-id2label.json""" elif task_name.startswith("""ade20k_""" ): _lowerCAmelCase : Any = 151 _lowerCAmelCase : Optional[Any] = 512 _lowerCAmelCase : Tuple = """ade20k-id2label.json""" _lowerCAmelCase : List[str] = True elif task_name.startswith("""voc_""" ): _lowerCAmelCase : Optional[int] = 21 _lowerCAmelCase : Union[str, Any] = 512 _lowerCAmelCase : str = """pascal-voc-id2label.json""" _lowerCAmelCase : List[Any] = True # orig_config _lowerCAmelCase : List[Any] = load_orig_config_file(_lowerCamelCase ) assert getattr(_lowerCamelCase ,"""model.classification.name""" ,-1 ) == "mobilevit_v2", "Invalid model" _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""model.classification.mitv2.width_multiplier""" ,1.0 ) assert ( getattr(_lowerCamelCase ,"""model.classification.mitv2.attn_norm_layer""" ,-1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" _lowerCAmelCase : Union[str, Any] = getattr(_lowerCamelCase ,"""model.classification.activation.name""" ,"""swish""" ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: _lowerCAmelCase : List[Any] = getattr(_lowerCamelCase ,"""model.segmentation.output_stride""" ,16 ) if "_deeplabv3" in task_name: _lowerCAmelCase : Optional[Any] = getattr(_lowerCamelCase ,"""model.segmentation.deeplabv3.aspp_rates""" ,[12, 24, 36] ) _lowerCAmelCase : List[str] = getattr(_lowerCamelCase ,"""model.segmentation.deeplabv3.aspp_out_channels""" ,512 ) _lowerCAmelCase : Optional[int] = getattr(_lowerCamelCase ,"""model.segmentation.deeplabv3.aspp_dropout""" ,0.1 ) # id2label _lowerCAmelCase : Dict = """huggingface/label-files""" _lowerCAmelCase : Optional[Any] = json.load(open(hf_hub_download(_lowerCamelCase ,_lowerCamelCase ,repo_type="""dataset""" ) ,"""r""" ) ) _lowerCAmelCase : Optional[Any] = {int(_lowerCamelCase ): v for k, v in idalabel.items()} _lowerCAmelCase : Optional[Any] = idalabel _lowerCAmelCase : str = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int] ,_lowerCamelCase : Union[str, Any] ,_lowerCamelCase : Any ) -> Optional[Any]: _lowerCAmelCase : List[Any] = dct.pop(_lowerCamelCase ) _lowerCAmelCase : List[Any] = val def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Tuple ,_lowerCamelCase : Optional[int]=False ) -> str: if base_model: _lowerCAmelCase : List[Any] = """""" else: _lowerCAmelCase : Optional[Any] = """mobilevitv2.""" _lowerCAmelCase : Dict = [] for k in state_dict.keys(): if k[:8] == "encoder.": _lowerCAmelCase : Tuple = k[8:] else: _lowerCAmelCase : Any = k if ".block." in k: _lowerCAmelCase : Any = k_new.replace(""".block.""" ,""".""" ) if ".conv." in k: _lowerCAmelCase : Any = k_new.replace(""".conv.""" ,""".convolution.""" ) if ".norm." in k: _lowerCAmelCase : List[Any] = k_new.replace(""".norm.""" ,""".normalization.""" ) if "conv_1." in k: _lowerCAmelCase : Dict = k_new.replace("""conv_1.""" ,f"{model_prefix}conv_stem." ) for i in [1, 2]: if f"layer_{i}." in k: _lowerCAmelCase : List[Any] = k_new.replace(f"layer_{i}." ,f"{model_prefix}encoder.layer.{i-1}.layer." ) if ".exp_1x1." in k: _lowerCAmelCase : List[Any] = k_new.replace(""".exp_1x1.""" ,""".expand_1x1.""" ) if ".red_1x1." in k: _lowerCAmelCase : str = k_new.replace(""".red_1x1.""" ,""".reduce_1x1.""" ) for i in [3, 4, 5]: if f"layer_{i}.0." in k: _lowerCAmelCase : List[Any] = k_new.replace(f"layer_{i}.0." ,f"{model_prefix}encoder.layer.{i-1}.downsampling_layer." ) if f"layer_{i}.1.local_rep.0." in k: _lowerCAmelCase : str = k_new.replace(f"layer_{i}.1.local_rep.0." ,f"{model_prefix}encoder.layer.{i-1}.conv_kxk." ) if f"layer_{i}.1.local_rep.1." in k: _lowerCAmelCase : Dict = k_new.replace(f"layer_{i}.1.local_rep.1." ,f"{model_prefix}encoder.layer.{i-1}.conv_1x1." ) for i in [3, 4, 5]: if i == 3: _lowerCAmelCase : Any = [0, 1] elif i == 4: _lowerCAmelCase : Dict = [0, 1, 2, 3] elif i == 5: _lowerCAmelCase : Optional[Any] = [0, 1, 2] for j in j_in: if f"layer_{i}.1.global_rep.{j}." in k: _lowerCAmelCase : List[Any] = k_new.replace( f"layer_{i}.1.global_rep.{j}." ,f"{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}." ) if f"layer_{i}.1.global_rep.{j+1}." in k: _lowerCAmelCase : Union[str, Any] = k_new.replace( f"layer_{i}.1.global_rep.{j+1}." ,f"{model_prefix}encoder.layer.{i-1}.layernorm." ) if f"layer_{i}.1.conv_proj." in k: _lowerCAmelCase : Any = k_new.replace(f"layer_{i}.1.conv_proj." ,f"{model_prefix}encoder.layer.{i-1}.conv_projection." ) if "pre_norm_attn.0." in k: _lowerCAmelCase : int = k_new.replace("""pre_norm_attn.0.""" ,"""layernorm_before.""" ) if "pre_norm_attn.1." in k: _lowerCAmelCase : int = k_new.replace("""pre_norm_attn.1.""" ,"""attention.""" ) if "pre_norm_ffn.0." in k: _lowerCAmelCase : int = k_new.replace("""pre_norm_ffn.0.""" ,"""layernorm_after.""" ) if "pre_norm_ffn.1." in k: _lowerCAmelCase : Optional[Any] = k_new.replace("""pre_norm_ffn.1.""" ,"""ffn.conv1.""" ) if "pre_norm_ffn.3." in k: _lowerCAmelCase : Any = k_new.replace("""pre_norm_ffn.3.""" ,"""ffn.conv2.""" ) if "classifier.1." in k: _lowerCAmelCase : Optional[Any] = k_new.replace("""classifier.1.""" ,"""classifier.""" ) if "seg_head." in k: _lowerCAmelCase : List[Any] = k_new.replace("""seg_head.""" ,"""segmentation_head.""" ) if ".aspp_layer." in k: _lowerCAmelCase : int = k_new.replace(""".aspp_layer.""" ,""".""" ) if ".aspp_pool." in k: _lowerCAmelCase : Dict = k_new.replace(""".aspp_pool.""" ,""".""" ) rename_keys.append((k, k_new) ) return rename_keys def SCREAMING_SNAKE_CASE ( _lowerCamelCase : str ) -> Tuple: _lowerCAmelCase : List[str] = [] for k in state_dict.keys(): if k.startswith("""seg_head.aux_head.""" ): keys_to_ignore.append(_lowerCamelCase ) for k in keys_to_ignore: state_dict.pop(_lowerCamelCase ,_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( ) -> List[str]: _lowerCAmelCase : List[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" _lowerCAmelCase : List[Any] = Image.open(requests.get(_lowerCamelCase ,stream=_lowerCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE ( _lowerCamelCase : List[str] ,_lowerCamelCase : Optional[Any] ,_lowerCamelCase : Dict ,_lowerCamelCase : Any ) -> Optional[Any]: _lowerCAmelCase : List[Any] = get_mobilevitva_config(_lowerCamelCase ,_lowerCamelCase ) # load original state_dict _lowerCAmelCase : Tuple = torch.load(_lowerCamelCase ,map_location="""cpu""" ) # load huggingface model if task_name.startswith("""ade20k_""" ) or task_name.startswith("""voc_""" ): _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(_lowerCamelCase ).eval() _lowerCAmelCase : int = False else: _lowerCAmelCase : Any = MobileViTVaForImageClassification(_lowerCamelCase ).eval() _lowerCAmelCase : Any = False # remove and rename some keys of load the original model _lowerCAmelCase : List[Any] = checkpoint remove_unused_keys(_lowerCamelCase ) _lowerCAmelCase : Tuple = create_rename_keys(_lowerCamelCase ,base_model=_lowerCamelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) # load modified state_dict model.load_state_dict(_lowerCamelCase ) # Check outputs on an image, prepared by MobileViTImageProcessor _lowerCAmelCase : str = MobileViTImageProcessor(crop_size=config.image_size ,size=config.image_size + 32 ) _lowerCAmelCase : Union[str, Any] = image_processor(images=prepare_img() ,return_tensors="""pt""" ) _lowerCAmelCase : Tuple = model(**_lowerCamelCase ) # verify classification model if task_name.startswith("""imagenet""" ): _lowerCAmelCase : List[str] = outputs.logits _lowerCAmelCase : List[Any] = logits.argmax(-1 ).item() print("""Predicted class:""" ,model.config.idalabel[predicted_class_idx] ) if task_name.startswith("""imagenet1k_256""" ) and config.width_multiplier == 1.0: # expected_logits for base variant _lowerCAmelCase : Tuple = torch.tensor([-1.6336e00, -7.3204e-02, -5.1883e-01] ) assert torch.allclose(logits[0, :3] ,_lowerCamelCase ,atol=1e-4 ) Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) print(f"Saving model {task_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_lowerCamelCase ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": _a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='imagenet1k_256', type=str, help=( 'Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . ' '\n Classification (ImageNet-1k)\n - MobileViTV2 (256x256) : imagenet1k_256\n - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384\n - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) :\n imagenet21k_to_1k_256\n - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on\n ImageNet-1k 384x384) : imagenet21k_to_1k_384\n Segmentation\n - ADE20K Dataset : ade20k_deeplabv3\n - Pascal VOC 2012 Dataset: voc_deeplabv3\n ' ), choices=[ 'imagenet1k_256', 'imagenet1k_384', 'imagenet21k_to_1k_256', 'imagenet21k_to_1k_384', 'ade20k_deeplabv3', 'voc_deeplabv3', ], ) parser.add_argument( '--orig_checkpoint_path', required=True, type=str, help='Path to the original state dict (.pt file).' ) parser.add_argument('--orig_config_path', required=True, type=str, help='Path to the original config file.') parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) _a : List[Any] = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
126
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : List[Any] = logging.get_logger(__name__) class __A ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Tuple = "maskformer-swin" _UpperCamelCase : Union[str, Any] = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , a__=224 , a__=4 , a__=3 , a__=96 , a__=[2, 2, 6, 2] , a__=[3, 6, 12, 24] , a__=7 , a__=4.0 , a__=True , a__=0.0 , a__=0.0 , a__=0.1 , a__="gelu" , a__=False , a__=0.0_2 , a__=1e-5 , a__=None , a__=None , **a__ , ): super().__init__(**a__ ) _lowerCAmelCase : Dict = image_size _lowerCAmelCase : List[str] = patch_size _lowerCAmelCase : Any = num_channels _lowerCAmelCase : int = embed_dim _lowerCAmelCase : Optional[Any] = depths _lowerCAmelCase : List[str] = len(a__ ) _lowerCAmelCase : List[Any] = num_heads _lowerCAmelCase : Tuple = window_size _lowerCAmelCase : List[Any] = mlp_ratio _lowerCAmelCase : Optional[Any] = qkv_bias _lowerCAmelCase : int = hidden_dropout_prob _lowerCAmelCase : Union[str, Any] = attention_probs_dropout_prob _lowerCAmelCase : Any = drop_path_rate _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : Tuple = use_absolute_embeddings _lowerCAmelCase : str = layer_norm_eps _lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowerCAmelCase : Union[str, Any] = int(embed_dim * 2 ** (len(a__ ) - 1) ) _lowerCAmelCase : int = ["""stem"""] + [F"stage{idx}" for idx in range(1 , len(a__ ) + 1 )] _lowerCAmelCase , _lowerCAmelCase : int = get_aligned_output_features_output_indices( out_features=a__ , out_indices=a__ , stage_names=self.stage_names )
126
1
from ...processing_utils import ProcessorMixin class UpperCAmelCase_ ( a): lowerCamelCase__ = 'SpeechT5FeatureExtractor' lowerCamelCase__ = 'SpeechT5Tokenizer' def __init__( self, __a, __a): '''simple docstring''' super().__init__(__a, __a) def __call__( self, *__a, **__a): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = kwargs.pop("audio", __a) _lowerCAmelCase : Dict = kwargs.pop("text", __a) _lowerCAmelCase : Dict = kwargs.pop("text_target", __a) _lowerCAmelCase : Union[str, Any] = kwargs.pop("audio_target", __a) _lowerCAmelCase : Any = kwargs.pop("sampling_rate", __a) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?") if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?") if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.") if audio is not None: _lowerCAmelCase : Tuple = self.feature_extractor(__a, *__a, sampling_rate=__a, **__a) elif text is not None: _lowerCAmelCase : List[Any] = self.tokenizer(__a, **__a) else: _lowerCAmelCase : Dict = None if audio_target is not None: _lowerCAmelCase : Union[str, Any] = self.feature_extractor(audio_target=__a, *__a, sampling_rate=__a, **__a) _lowerCAmelCase : Optional[int] = targets["input_values"] elif text_target is not None: _lowerCAmelCase : List[Any] = self.tokenizer(__a, **__a) _lowerCAmelCase : Union[str, Any] = targets["input_ids"] else: _lowerCAmelCase : Union[str, Any] = None if inputs is None: return targets if targets is not None: _lowerCAmelCase : Any = labels _lowerCAmelCase : List[Any] = targets.get("attention_mask") if decoder_attention_mask is not None: _lowerCAmelCase : Tuple = decoder_attention_mask return inputs def snake_case__ ( self, *__a, **__a): '''simple docstring''' _lowerCAmelCase : List[str] = kwargs.pop("input_values", __a) _lowerCAmelCase : int = kwargs.pop("input_ids", __a) _lowerCAmelCase : List[Any] = kwargs.pop("labels", __a) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs.") if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.") if input_values is not None: _lowerCAmelCase : List[str] = self.feature_extractor.pad(__a, *__a, **__a) elif input_ids is not None: _lowerCAmelCase : Optional[Any] = self.tokenizer.pad(__a, **__a) else: _lowerCAmelCase : List[Any] = None if labels is not None: if "input_ids" in labels or (isinstance(__a, __a) and "input_ids" in labels[0]): _lowerCAmelCase : str = self.tokenizer.pad(__a, **__a) _lowerCAmelCase : str = targets["input_ids"] else: _lowerCAmelCase : Union[str, Any] = self.feature_extractor.feature_size _lowerCAmelCase : str = self.feature_extractor.num_mel_bins _lowerCAmelCase : str = self.feature_extractor.pad(__a, *__a, **__a) _lowerCAmelCase : List[Any] = feature_size_hack _lowerCAmelCase : str = targets["input_values"] else: _lowerCAmelCase : Optional[Any] = None if inputs is None: return targets if targets is not None: _lowerCAmelCase : str = labels _lowerCAmelCase : List[str] = targets.get("attention_mask") if decoder_attention_mask is not None: _lowerCAmelCase : Any = decoder_attention_mask return inputs def snake_case__ ( self, *__a, **__a): '''simple docstring''' return self.tokenizer.batch_decode(*__a, **__a) def snake_case__ ( self, *__a, **__a): '''simple docstring''' return self.tokenizer.decode(*__a, **__a)
36
import argparse import copy def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : int = {} with open(_lowerCamelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _lowerCAmelCase : Tuple = [] _list.append([line.split()[1], line.split()[2]] ) _lowerCAmelCase : Any = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _lowerCAmelCase : str = [] _list.append([line.split()[0], line.split()[2]] ) _lowerCAmelCase : Any = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' with open(_lowerCamelCase ) as f: _lowerCAmelCase : str = f.read(1 ) _lowerCAmelCase : str = start_node _lowerCAmelCase : List[str] = [] _lowerCAmelCase : Any = start_node _lowerCAmelCase : str = 0 while visiting not in first_solution: _lowerCAmelCase : Dict = 10_000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(_lowerCamelCase ) and k[0] not in first_solution: _lowerCAmelCase : List[str] = k[1] _lowerCAmelCase : List[Any] = k[0] first_solution.append(_lowerCamelCase ) _lowerCAmelCase : Optional[int] = distance_of_first_solution + int(_lowerCamelCase ) _lowerCAmelCase : str = best_node first_solution.append(_lowerCamelCase ) _lowerCAmelCase : Union[str, Any] = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _lowerCAmelCase : Tuple = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10_000 ) return first_solution, distance_of_first_solution def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Tuple = [] for n in solution[1:-1]: _lowerCAmelCase : Dict = solution.index(_lowerCamelCase ) for kn in solution[1:-1]: _lowerCAmelCase : Dict = solution.index(_lowerCamelCase ) if n == kn: continue _lowerCAmelCase : Optional[int] = copy.deepcopy(_lowerCamelCase ) _lowerCAmelCase : int = kn _lowerCAmelCase : Dict = n _lowerCAmelCase : Optional[int] = 0 for k in _tmp[:-1]: _lowerCAmelCase : str = _tmp[_tmp.index(_lowerCamelCase ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _lowerCAmelCase : Optional[Any] = distance + int(i[1] ) _tmp.append(_lowerCamelCase ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _lowerCAmelCase : List[Any] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda _lowerCamelCase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : List[str] = 1 _lowerCAmelCase : int = first_solution _lowerCAmelCase : Tuple = [] _lowerCAmelCase : Tuple = distance_of_first_solution _lowerCAmelCase : Optional[int] = solution while count <= iters: _lowerCAmelCase : int = find_neighborhood(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : Tuple = 0 _lowerCAmelCase : Dict = neighborhood[index_of_best_solution] _lowerCAmelCase : int = len(_lowerCamelCase ) - 1 _lowerCAmelCase : Union[str, Any] = False while not found: _lowerCAmelCase : Tuple = 0 while i < len(_lowerCamelCase ): if best_solution[i] != solution[i]: _lowerCAmelCase : str = best_solution[i] _lowerCAmelCase : Tuple = solution[i] break _lowerCAmelCase : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _lowerCAmelCase : Optional[int] = True _lowerCAmelCase : Optional[Any] = best_solution[:-1] _lowerCAmelCase : Tuple = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _lowerCAmelCase : Union[str, Any] = cost _lowerCAmelCase : List[Any] = solution else: _lowerCAmelCase : Optional[Any] = index_of_best_solution + 1 _lowerCAmelCase : Optional[Any] = neighborhood[index_of_best_solution] if len(_lowerCamelCase ) >= size: tabu_list.pop(0 ) _lowerCAmelCase : int = count + 1 return best_solution_ever, best_cost def A ( _lowerCamelCase=None ): '''simple docstring''' _lowerCAmelCase : int = generate_neighbours(args.File ) _lowerCAmelCase , _lowerCAmelCase : List[str] = generate_first_solution( args.File , _lowerCamelCase ) _lowerCAmelCase , _lowerCAmelCase : Any = tabu_search( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , args.Iterations , args.Size , ) print(F"Best solution: {best_sol}, with total distance: {best_cost}." ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
36
1
def snake_case_ (__A : int = 1_0_0 ) -> int: __lowerCAmelCase : Optional[Any] = set() __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : Dict = n + 1 # maximum limit for a in range(2 , __A ): for b in range(2 , __A ): __lowerCAmelCase : Optional[int] = a**b # calculates the current power collect_powers.add(__A ) # adds the result to the set return len(__A ) if __name__ == "__main__": print("""Number of terms """, solution(int(str(input()).strip())))
139
import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowerCamelCase : str =DebertaTokenizer lowerCamelCase : Dict =True lowerCamelCase : List[str] =DebertaTokenizerFast def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Union[str, Any]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowerCAmelCase : Optional[int] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """[UNK]""", ] __lowerCAmelCase : Optional[int] = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) __lowerCAmelCase : List[str] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __lowerCAmelCase : Any = {"""unk_token""": """[UNK]"""} __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCAmelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowerCAmelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , **lowerCAmelCase : Tuple ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase : int ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Dict = """lower newer""" __lowerCAmelCase : Union[str, Any] = """lower newer""" return input_text, output_text def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.get_tokenizer() __lowerCAmelCase : int = """lower newer""" __lowerCAmelCase : Any = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] __lowerCAmelCase : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) __lowerCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __lowerCAmelCase : Union[str, Any] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase ) , lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Any = tokenizer("""Hello""" , """World""" ) __lowerCAmelCase : Tuple = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd["""token_type_ids"""] , lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = self.tokenizer_class.from_pretrained("""microsoft/deberta-base""" ) __lowerCAmelCase : Optional[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=lowerCAmelCase ) __lowerCAmelCase : List[Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=lowerCAmelCase ) __lowerCAmelCase : Dict = tokenizer.encode( """sequence builders""" , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) __lowerCAmelCase : Dict = tokenizer.encode( """sequence builders""" , """multi-sequence build""" , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) __lowerCAmelCase : List[str] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase ) __lowerCAmelCase : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase , lowerCAmelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: __lowerCAmelCase : int = tokenizer_class.from_pretrained("""microsoft/deberta-base""" ) __lowerCAmelCase : Optional[Any] = [ """ALBERT: A Lite BERT for Self-supervised Learning of Language Representations""", """ALBERT incorporates two parameter reduction techniques""", """The first one is a factorized embedding parameterization. By decomposing the large vocabulary""" """ embedding matrix into two small matrices, we separate the size of the hidden layers from the size of""" """ vocabulary embedding.""", ] __lowerCAmelCase : List[str] = tokenizer(lowerCAmelCase , padding=lowerCAmelCase ) __lowerCAmelCase : List[str] = [tokenizer.decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) for seq in encoding["""input_ids"""]] # fmt: off __lowerCAmelCase : Any = { """input_ids""": [ [1, 21_18, 1_11_26, 5_65, 35, 83, 2_51_91, 1_63, 1_88_54, 13, 1_21_56, 12, 1_61_01, 2_53_76, 1_38_07, 9, 2_22_05, 2_78_93, 16_35, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 21_18, 1_11_26, 5_65, 2_45_36, 80, 4_37_97, 48_78, 73_73, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1_33, 78, 65, 16, 10, 37_24, 15_38, 3_31_83, 1_13_03, 4_37_97, 19_38, 4, 8_70, 2_41_65, 2_91_05, 5, 7_39, 3_26_44, 3_31_83, 1_13_03, 3_61_73, 88, 80, 6_50, 78_21, 4_59_40, 6, 52, 25_59, 5, 18_36, 9, 5, 73_97, 1_31_71, 31, 5, 18_36, 9, 3_26_44, 3_31_83, 1_13_03, 4, 2] ], """token_type_ids""": [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on __lowerCAmelCase : int = [ """ALBERT: A Lite BERT for Self-supervised Learning of Language Representations""", """ALBERT incorporates two parameter reduction techniques""", """The first one is a factorized embedding parameterization. By decomposing the large vocabulary""" """ embedding matrix into two small matrices, we separate the size of the hidden layers from the size of""" """ vocabulary embedding.""", ] self.assertDictEqual(encoding.data , lowerCAmelCase ) for expected, decoded in zip(lowerCAmelCase , lowerCAmelCase ): self.assertEqual(lowerCAmelCase , lowerCAmelCase )
139
1
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _snake_case : Dict = "src/transformers" # This is to make sure the transformers module imported is the one in the repo. _snake_case : Optional[int] = direct_transformers_import(PATH_TO_TRANSFORMERS) _snake_case : Optional[int] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` _snake_case : int = re.compile(R"\[(.+?)\]\((https://huggingface\.co/.+?)\)") _snake_case : Optional[int] = { "DecisionTransformerConfig", "EncoderDecoderConfig", "MusicgenConfig", "RagConfig", "SpeechEncoderDecoderConfig", "TimmBackboneConfig", "VisionEncoderDecoderConfig", "VisionTextDualEncoderConfig", "LlamaConfig", } def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = None # source code of `config_class` __snake_case : Dict = inspect.getsource(__lowerCamelCase ) __snake_case : str = _re_checkpoint.findall(__lowerCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): __snake_case : Optional[Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __snake_case : List[Any] = F'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: __snake_case : List[Any] = ckpt_name break return checkpoint def lowerCAmelCase_ ( ): __snake_case : str = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __snake_case : Optional[int] = get_checkpoint_from_config_class(__lowerCamelCase ) __snake_case : str = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__lowerCamelCase ) if len(__lowerCamelCase ) > 0: __snake_case : int = "\n".join(sorted(__lowerCamelCase ) ) raise ValueError(F'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
123
import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) _snake_case : Any = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation="relu") ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation="relu")) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation="relu")) classifier.add(layers.Dense(units=1, activation="sigmoid")) # Compiling the CNN classifier.compile( optimizer="adam", loss="binary_crossentropy", metrics=["accuracy"] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') _snake_case : int = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) _snake_case : Optional[Any] = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) _snake_case : List[str] = train_datagen.flow_from_directory( "dataset/training_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) _snake_case : Any = test_datagen.flow_from_directory( "dataset/test_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save("cnn.h5") # Part 3 - Making new predictions _snake_case : Optional[Any] = tf.keras.preprocessing.image.load_img( "dataset/single_prediction/image.png", target_size=(64, 64) ) _snake_case : int = tf.keras.preprocessing.image.img_to_array(test_image) _snake_case : Tuple = np.expand_dims(test_image, axis=0) _snake_case : Any = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: _snake_case : Any = "Normal" if result[0][0] == 1: _snake_case : List[str] = "Abnormality detected"
123
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def _lowercase ( __lowerCAmelCase ) -> str: SCREAMING_SNAKE_CASE__ : Union[str, Any] = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __a (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :Optional[Any] = StableDiffusionLatentUpscalePipeline _SCREAMING_SNAKE_CASE :Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """height""", """width""", """cross_attention_kwargs""", """negative_prompt_embeds""", """prompt_embeds""", } _SCREAMING_SNAKE_CASE :str = PipelineTesterMixin.required_optional_params - {"""num_images_per_prompt"""} _SCREAMING_SNAKE_CASE :Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _SCREAMING_SNAKE_CASE :int = frozenset( []) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _SCREAMING_SNAKE_CASE :List[Any] = frozenset([]) _SCREAMING_SNAKE_CASE :List[str] = True @property def _a ( self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = 1 SCREAMING_SNAKE_CASE__ : Tuple = 4 SCREAMING_SNAKE_CASE__ : List[str] = (16, 16) SCREAMING_SNAKE_CASE__ : Dict = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_a ) return image def _a ( self ) -> Any: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = UNetaDConditionModel( act_fn="""gelu""" , attention_head_dim=8 , norm_num_groups=_a , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=160 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( """KDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", ) , in_channels=8 , mid_block_type=_a , only_cross_attention=_a , out_channels=5 , resnet_time_scale_shift="""scale_shift""" , time_embedding_type="""fourier""" , timestep_post_act="""gelu""" , up_block_types=("""KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KUpBlock2D""") , ) SCREAMING_SNAKE_CASE__ : Any = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", ] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) SCREAMING_SNAKE_CASE__ : Tuple = EulerDiscreteScheduler(prediction_type="""sample""" ) SCREAMING_SNAKE_CASE__ : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""quick_gelu""" , projection_dim=512 , ) SCREAMING_SNAKE_CASE__ : Optional[int] = CLIPTextModel(_a ) SCREAMING_SNAKE_CASE__ : Tuple = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) SCREAMING_SNAKE_CASE__ : Any = { """unet""": model.eval(), """vae""": vae.eval(), """scheduler""": scheduler, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def _a ( self , _a , _a=0 ) -> Tuple: """simple docstring""" if str(_a ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ : List[str] = torch.manual_seed(_a ) else: SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.Generator(device=_a ).manual_seed(_a ) SCREAMING_SNAKE_CASE__ : Tuple = { """prompt""": """A painting of a squirrel eating a burger""", """image""": self.dummy_image.cpu(), """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = """cpu""" SCREAMING_SNAKE_CASE__ : List[str] = self.get_dummy_components() SCREAMING_SNAKE_CASE__ : Tuple = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_dummy_inputs(_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = pipe(**_a ).images SCREAMING_SNAKE_CASE__ : str = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 256, 256, 3) ) SCREAMING_SNAKE_CASE__ : List[str] = np.array( [0.47_222_412, 0.41_921_633, 0.44_717_434, 0.46_874_192, 0.42_588_258, 0.46_150_726, 0.4_677_534, 0.45_583_832, 0.48_579_055] ) SCREAMING_SNAKE_CASE__ : List[Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a , 1E-3 ) def _a ( self ) -> Tuple: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=7E-3 ) def _a ( self ) -> List[str]: """simple docstring""" super().test_cpu_offload_forward_pass(expected_max_diff=3E-3 ) def _a ( self ) -> List[str]: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def _a ( self ) -> Dict: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=7E-3 ) def _a ( self ) -> Any: """simple docstring""" super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3E-3 ) def _a ( self ) -> str: """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def _a ( self ) -> Optional[int]: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = [ """DDIMScheduler""", """DDPMScheduler""", """PNDMScheduler""", """HeunDiscreteScheduler""", """EulerAncestralDiscreteScheduler""", """KDPM2DiscreteScheduler""", """KDPM2AncestralDiscreteScheduler""", """DPMSolverSDEScheduler""", ] SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_dummy_components() SCREAMING_SNAKE_CASE__ : List[str] = self.pipeline_class(**_a ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : int = self.get_dummy_inputs(_a ) SCREAMING_SNAKE_CASE__ : List[Any] = 2 SCREAMING_SNAKE_CASE__ : int = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue SCREAMING_SNAKE_CASE__ : Optional[int] = getattr(_a , scheduler_enum.name ) SCREAMING_SNAKE_CASE__ : Any = scheduler_cls.from_config(pipe.scheduler.config ) SCREAMING_SNAKE_CASE__ : List[str] = pipe(**_a )[0] outputs.append(_a ) assert check_same_shape(_a ) @require_torch_gpu @slow class __a (unittest.TestCase): '''simple docstring''' def _a ( self ) -> Any: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE__ : List[str] = StableDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" , torch_dtype=torch.floataa ) pipe.to("""cuda""" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) SCREAMING_SNAKE_CASE__ : Dict = """a photo of an astronaut high resolution, unreal engine, ultra realistic""" SCREAMING_SNAKE_CASE__ : Dict = pipe(_a , generator=_a , output_type="""latent""" ).images SCREAMING_SNAKE_CASE__ : Tuple = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type="""np""" , ).images[0] SCREAMING_SNAKE_CASE__ : List[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy""" ) assert np.abs((expected_image - image).mean() ) < 5E-2 def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE__ : str = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) SCREAMING_SNAKE_CASE__ : Any = """the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas""" SCREAMING_SNAKE_CASE__ : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png""" ) SCREAMING_SNAKE_CASE__ : Dict = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type="""np""" , ).images[0] SCREAMING_SNAKE_CASE__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy""" ) assert np.abs((expected_image - image).max() ) < 5E-2
350
"""simple docstring""" def _lowercase ( __lowerCAmelCase ) -> bool: if number < 0: raise ValueError("""number must not be negative""" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
56
0
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class __lowercase ( _lowerCAmelCase ): """simple docstring""" UpperCamelCase : List[str] = (UnCLIPScheduler,) def __A ( self , **A ) -> Any: '''simple docstring''' lowerCamelCase = { "num_train_timesteps": 10_00, "variance_type": "fixed_small_log", "clip_sample": True, "clip_sample_range": 1.0, "prediction_type": "epsilon", } config.update(**A ) return config def __A ( self ) -> Any: '''simple docstring''' for timesteps in [1, 5, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=A ) def __A ( self ) -> List[str]: '''simple docstring''' for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=A ) def __A ( self ) -> Dict: '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=A ) def __A ( self ) -> Optional[int]: '''simple docstring''' for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=A ) def __A ( self ) -> Tuple: '''simple docstring''' for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=A ) def __A ( self ) -> Dict: '''simple docstring''' for time_step in [0, 5_00, 9_99]: for prev_timestep in [None, 5, 1_00, 2_50, 5_00, 7_50]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=A , prev_timestep=A ) def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config(variance_type="""fixed_small_log""" ) lowerCamelCase = scheduler_class(**A ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0_0_0_0e-1_0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 ) - 0.0549625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 ) - 0.9994987 ) ) < 1e-5 def __A ( self ) -> Tuple: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config(variance_type="""learned_range""" ) lowerCamelCase = scheduler_class(**A ) lowerCamelCase = 0.5 assert scheduler._get_variance(1 , predicted_variance=A ) - -10.1712790 < 1e-5 assert scheduler._get_variance(4_87 , predicted_variance=A ) - -5.7998052 < 1e-5 assert scheduler._get_variance(9_99 , predicted_variance=A ) - -0.0010011 < 1e-5 def __A ( self ) -> Dict: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config() lowerCamelCase = scheduler_class(**A ) lowerCamelCase = scheduler.timesteps lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter lowerCamelCase = torch.manual_seed(0 ) for i, t in enumerate(A ): # 1. predict noise residual lowerCamelCase = model(A , A ) # 2. predict previous mean of sample x_t-1 lowerCamelCase = scheduler.step(A , A , A , generator=A ).prev_sample lowerCamelCase = pred_prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 2_52.2_68_24_95 ) < 1e-2 assert abs(result_mean.item() - 0.3284743 ) < 1e-3 def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config() lowerCamelCase = scheduler_class(**A ) scheduler.set_timesteps(25 ) lowerCamelCase = scheduler.timesteps lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter lowerCamelCase = torch.manual_seed(0 ) for i, t in enumerate(A ): # 1. predict noise residual lowerCamelCase = model(A , A ) if i + 1 == timesteps.shape[0]: lowerCamelCase = None else: lowerCamelCase = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 lowerCamelCase = scheduler.step( A , A , A , prev_timestep=A , generator=A ).prev_sample lowerCamelCase = pred_prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 2_58.2_04_49_83 ) < 1e-2 assert abs(result_mean.item() - 0.3362038 ) < 1e-3 def __A ( self ) -> Tuple: '''simple docstring''' pass def __A ( self ) -> int: '''simple docstring''' pass
252
'''simple docstring''' import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def UpperCAmelCase_ ( __lowerCamelCase : Optional[int] ,__lowerCamelCase : Any ,__lowerCamelCase : Any ,__lowerCamelCase : List[str] ,__lowerCamelCase : Dict=True ,__lowerCamelCase : List[Any]="pt" ): lowercase_ :Dict = {"add_prefix_space": True} if isinstance(__lowerCamelCase ,__lowerCamelCase ) and not line.startswith(" " ) else {} lowercase_ :str = padding_side return tokenizer( [line] ,max_length=__lowerCamelCase ,padding="max_length" if pad_to_max_length else None ,truncation=__lowerCamelCase ,return_tensors=__lowerCamelCase ,add_special_tokens=__lowerCamelCase ,**__lowerCamelCase ,) def UpperCAmelCase_ ( __lowerCamelCase : List[str] ,__lowerCamelCase : Dict ,__lowerCamelCase : str=None ,): lowercase_ :Optional[int] = input_ids.ne(__lowerCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class a_ ( _lowerCAmelCase ): def __init__( self : Optional[int] , lowercase : Any , lowercase : Tuple , lowercase : Union[str, Any] , lowercase : Tuple , lowercase : str="train" , lowercase : Dict=None , lowercase : Tuple=None , lowercase : List[str]=None , lowercase : int="" , ): """simple docstring""" super().__init__() lowercase_ :List[Any] = Path(lowercase ).joinpath(type_path + ".source" ) lowercase_ :Dict = Path(lowercase ).joinpath(type_path + ".target" ) lowercase_ :Optional[int] = self.get_char_lens(self.src_file ) lowercase_ :List[str] = max_source_length lowercase_ :str = max_target_length assert min(self.src_lens ) > 0, F'found empty line in {self.src_file}' lowercase_ :int = tokenizer lowercase_ :Dict = prefix if n_obs is not None: lowercase_ :Union[str, Any] = self.src_lens[:n_obs] lowercase_ :Optional[int] = src_lang lowercase_ :str = tgt_lang def __len__( self : Tuple ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : str , lowercase : Dict ): """simple docstring""" lowercase_ :Tuple = index + 1 # linecache starts at 1 lowercase_ :Optional[Any] = self.prefix + linecache.getline(str(self.src_file ) , lowercase ).rstrip("\n" ) lowercase_ :List[str] = linecache.getline(str(self.tgt_file ) , lowercase ).rstrip("\n" ) assert source_line, F'empty source line for index {index}' assert tgt_line, F'empty tgt line for index {index}' # Need to add eos token manually for T5 if isinstance(self.tokenizer , lowercase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right lowercase_ :List[str] = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , lowercase ) else self.tokenizer ) lowercase_ :int = self.tokenizer.generator if isinstance(self.tokenizer , lowercase ) else self.tokenizer lowercase_ :List[str] = encode_line(lowercase , lowercase , self.max_source_length , "right" ) lowercase_ :Any = encode_line(lowercase , lowercase , self.max_target_length , "right" ) lowercase_ :Dict = source_inputs["input_ids"].squeeze() lowercase_ :Tuple = target_inputs["input_ids"].squeeze() lowercase_ :Optional[int] = source_inputs["attention_mask"].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def lowercase__ ( lowercase : Union[str, Any] ): """simple docstring""" return [len(lowercase ) for x in Path(lowercase ).open().readlines()] def lowercase__ ( self : str , lowercase : List[Any] ): """simple docstring""" lowercase_ :Optional[int] = torch.stack([x["input_ids"] for x in batch] ) lowercase_ :Dict = torch.stack([x["attention_mask"] for x in batch] ) lowercase_ :List[str] = torch.stack([x["decoder_input_ids"] for x in batch] ) lowercase_ :Any = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , lowercase ) else self.tokenizer.pad_token_id ) lowercase_ :str = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , lowercase ) else self.tokenizer.pad_token_id ) lowercase_ :Union[str, Any] = trim_batch(lowercase , lowercase ) lowercase_ , lowercase_ :Optional[Any] = trim_batch(lowercase , lowercase , attention_mask=lowercase ) lowercase_ :Tuple = { "input_ids": source_ids, "attention_mask": source_mask, "decoder_input_ids": y, } return batch lowerCAmelCase : List[str] =getLogger(__name__) def UpperCAmelCase_ ( __lowerCamelCase : List[List] ): return list(itertools.chain.from_iterable(__lowerCamelCase ) ) def UpperCAmelCase_ ( __lowerCamelCase : str ): lowercase_ :List[str] = get_git_info() save_json(__lowerCamelCase ,os.path.join(__lowerCamelCase ,"git_log.json" ) ) def UpperCAmelCase_ ( __lowerCamelCase : Any ,__lowerCamelCase : Any ,__lowerCamelCase : List[Any]=4 ,**__lowerCamelCase : List[str] ): with open(__lowerCamelCase ,"w" ) as f: json.dump(__lowerCamelCase ,__lowerCamelCase ,indent=__lowerCamelCase ,**__lowerCamelCase ) def UpperCAmelCase_ ( __lowerCamelCase : Tuple ): with open(__lowerCamelCase ) as f: return json.load(__lowerCamelCase ) def UpperCAmelCase_ ( ): lowercase_ :Dict = git.Repo(search_parent_directories=__lowerCamelCase ) lowercase_ :List[str] = { "repo_id": str(__lowerCamelCase ), "repo_sha": str(repo.head.object.hexsha ), "repo_branch": str(repo.active_branch ), "hostname": str(socket.gethostname() ), } return repo_infos def UpperCAmelCase_ ( __lowerCamelCase : Callable ,__lowerCamelCase : Iterable ): return list(map(__lowerCamelCase ,__lowerCamelCase ) ) def UpperCAmelCase_ ( __lowerCamelCase : Union[str, Any] ,__lowerCamelCase : List[str] ): with open(__lowerCamelCase ,"wb" ) as f: return pickle.dump(__lowerCamelCase ,__lowerCamelCase ) def UpperCAmelCase_ ( __lowerCamelCase : str ): def remove_articles(__lowerCamelCase : Optional[int] ): return re.sub(r"\b(a|an|the)\b" ," " ,__lowerCamelCase ) def white_space_fix(__lowerCamelCase : Dict ): return " ".join(text.split() ) def remove_punc(__lowerCamelCase : Optional[Any] ): lowercase_ :Any = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__lowerCamelCase : List[Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__lowerCamelCase ) ) ) ) def UpperCAmelCase_ ( __lowerCamelCase : Optional[int] ,__lowerCamelCase : Optional[int] ): lowercase_ :Tuple = normalize_answer(__lowerCamelCase ).split() lowercase_ :Dict = normalize_answer(__lowerCamelCase ).split() lowercase_ :Tuple = Counter(__lowerCamelCase ) & Counter(__lowerCamelCase ) lowercase_ :Tuple = sum(common.values() ) if num_same == 0: return 0 lowercase_ :Union[str, Any] = 1.0 * num_same / len(__lowerCamelCase ) lowercase_ :List[Any] = 1.0 * num_same / len(__lowerCamelCase ) lowercase_ :Tuple = (2 * precision * recall) / (precision + recall) return fa def UpperCAmelCase_ ( __lowerCamelCase : int ,__lowerCamelCase : Union[str, Any] ): return normalize_answer(__lowerCamelCase ) == normalize_answer(__lowerCamelCase ) def UpperCAmelCase_ ( __lowerCamelCase : List[str] ,__lowerCamelCase : List[str] ): assert len(__lowerCamelCase ) == len(__lowerCamelCase ) lowercase_ :Any = 0 for hypo, pred in zip(__lowerCamelCase ,__lowerCamelCase ): em += exact_match_score(__lowerCamelCase ,__lowerCamelCase ) if len(__lowerCamelCase ) > 0: em /= len(__lowerCamelCase ) return {"em": em} def UpperCAmelCase_ ( __lowerCamelCase : str ): return model_prefix.startswith("rag" ) def UpperCAmelCase_ ( __lowerCamelCase : List[str] ,__lowerCamelCase : List[Any] ,__lowerCamelCase : int ): lowercase_ :Dict = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead lowercase_ :List[str] = "dropout_rate" for p in extra_params: if getattr(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ): if not hasattr(__lowerCamelCase ,__lowerCamelCase ) and not hasattr(__lowerCamelCase ,equivalent_param[p] ): logger.info("config doesn't have a `{}` attribute".format(__lowerCamelCase ) ) delattr(__lowerCamelCase ,__lowerCamelCase ) continue lowercase_ :List[Any] = p if hasattr(__lowerCamelCase ,__lowerCamelCase ) else equivalent_param[p] setattr(__lowerCamelCase ,__lowerCamelCase ,getattr(__lowerCamelCase ,__lowerCamelCase ) ) delattr(__lowerCamelCase ,__lowerCamelCase ) return hparams, config
223
0
from collections import namedtuple lowerCAmelCase_ = namedtuple('''from_to''', '''from_ to''') lowerCAmelCase_ = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.001, 1_0_0_0), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.00_454, 264.172), '''cubicyard''': from_to(0.76_455, 1.30_795), '''cubicfoot''': from_to(0.028, 35.3_147), '''cup''': from_to(0.000_236_588, 4_226.75), } def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + ''', '''.join(_UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + ''', '''.join(_UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
279
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder lowerCAmelCase_ = '''__DUMMY_TRANSFORMERS_USER__''' lowerCAmelCase_ = '''Dummy User''' lowerCAmelCase_ = '''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''' lowerCAmelCase_ = '''https://hub-ci.huggingface.co''' lowerCAmelCase_ = CI_HUB_ENDPOINT + '''/datasets/{repo_id}/resolve/{revision}/{path}''' lowerCAmelCase_ = CI_HUB_ENDPOINT + '''/{repo_id}/resolve/{revision}/{filename}''' lowerCAmelCase_ = Path('''~/.huggingface/hub_ci_token''').expanduser() @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase ) -> Optional[int]: """simple docstring""" monkeypatch.setattr( '''huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE''' , _UpperCamelCase ) @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase ) -> Optional[int]: """simple docstring""" monkeypatch.setattr('''datasets.config.HF_ENDPOINT''' , _UpperCamelCase ) monkeypatch.setattr('''datasets.config.HUB_DATASETS_URL''' , _UpperCamelCase ) @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase ) -> Any: """simple docstring""" monkeypatch.setattr('''huggingface_hub.hf_api.HfFolder.path_token''' , _UpperCamelCase ) @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" HfFolder.save_token(_UpperCamelCase ) yield HfFolder.delete_token() @pytest.fixture(scope='''session''' ) def lowerCamelCase_ ( ) -> List[str]: """simple docstring""" return HfApi(endpoint=_UpperCamelCase ) @pytest.fixture(scope='''session''' ) def lowerCamelCase_ ( _UpperCamelCase ) -> Dict: """simple docstring""" snake_case_ : Union[str, Any] = HfFolder.get_token() HfFolder.save_token(_UpperCamelCase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_UpperCamelCase ) @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase ) -> List[str]: """simple docstring""" def _cleanup_repo(_UpperCamelCase ): hf_api.delete_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' ) return _cleanup_repo @pytest.fixture def lowerCamelCase_ ( _UpperCamelCase ) -> List[str]: """simple docstring""" @contextmanager def _temporary_repo(_UpperCamelCase ): try: yield repo_id finally: cleanup_repo(_UpperCamelCase ) return _temporary_repo @pytest.fixture(scope='''session''' ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" snake_case_ : Optional[int] = f'''repo_txt_data-{int(time.time() * 10E3 )}''' snake_case_ : Any = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' , private=_UpperCamelCase ) hf_api.upload_file( token=_UpperCamelCase , path_or_fileobj=str(_UpperCamelCase ) , path_in_repo='''data/text_data.txt''' , repo_id=_UpperCamelCase , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" snake_case_ : int = f'''repo_zipped_txt_data-{int(time.time() * 10E3 )}''' snake_case_ : Tuple = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' , private=_UpperCamelCase ) hf_api.upload_file( token=_UpperCamelCase , path_or_fileobj=str(_UpperCamelCase ) , path_in_repo='''data.zip''' , repo_id=_UpperCamelCase , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" snake_case_ : List[str] = f'''repo_zipped_img_data-{int(time.time() * 10E3 )}''' snake_case_ : str = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' , private=_UpperCamelCase ) hf_api.upload_file( token=_UpperCamelCase , path_or_fileobj=str(_UpperCamelCase ) , path_in_repo='''data.zip''' , repo_id=_UpperCamelCase , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(_UpperCamelCase , token=_UpperCamelCase , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> int: """simple docstring""" return hf_private_dataset_repo_zipped_img_data_
279
1
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __lowerCAmelCase ( unittest.TestCase ): def snake_case ( self , _snake_case ): """simple docstring""" _lowerCAmelCase = 3 _lowerCAmelCase = 250 _lowerCAmelCase = ids_tensor((batch_size, length) , _snake_case ) _lowerCAmelCase = torch.ones((batch_size, length) , device=_snake_case , dtype=torch.float ) / length return input_ids, scores def snake_case ( self ): """simple docstring""" _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(5 ) _lowerCAmelCase = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = MaxLengthCriteria(max_length=10 ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(10 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase , _lowerCAmelCase = self._get_tensors(5 ) _lowerCAmelCase = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(_snake_case , _snake_case ) ) _lowerCAmelCase = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(_snake_case , _snake_case ) ) def snake_case ( self ): """simple docstring""" validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(_snake_case ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) _lowerCAmelCase = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(_snake_case ) , 1 )
82
from math import isqrt, loga def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , snake_case , snake_case ): _lowerCAmelCase = False return [i for i in range(2 , snake_case ) if is_prime[i]] def _UpperCAmelCase ( snake_case = 80_08_00 , snake_case = 80_08_00 ): """simple docstring""" _lowerCAmelCase = degree * loga(snake_case ) _lowerCAmelCase = int(snake_case ) _lowerCAmelCase = calculate_prime_numbers(snake_case ) _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = len(snake_case ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(f"{solution() = }")
82
1
'''simple docstring''' import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __snake_case =logging.get_logger(__name__) __snake_case ="""▁""" __snake_case ={"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} __snake_case ={ """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } __snake_case ={ """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } __snake_case ={ """ernie-m-base""": 514, """ernie-m-large""": 514, } __snake_case ={ """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : List[str] = ["input_ids"] lowerCamelCase : Tuple = VOCAB_FILES_NAMES lowerCamelCase : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Dict = RESOURCE_FILES_NAMES def __init__( self : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : Dict="utf8" , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : Optional[Any]="[SEP]" , UpperCAmelCase__ : List[Any]="[PAD]" , UpperCAmelCase__ : int="[CLS]" , UpperCAmelCase__ : Tuple="[MASK]" , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : Optional[int] , ) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , vocab_file=UpperCAmelCase__ , encoding=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) lowerCAmelCase = do_lower_case lowerCAmelCase = sentencepiece_model_ckpt lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowerCAmelCase = self.load_vocab(filepath=UpperCAmelCase__ ) else: lowerCAmelCase = {self.sp_model.id_to_piece(UpperCAmelCase__ ): id for id in range(self.sp_model.get_piece_size() )} lowerCAmelCase = {v: k for k, v in self.vocab.items()} def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : int ) -> str: if text is None: return None lowerCAmelCase = self.tokenize(UpperCAmelCase__ ) lowerCAmelCase , lowerCAmelCase = '', [] for i, ch in enumerate(UpperCAmelCase__ ): if ch in self.SP_CHAR_MAPPING: lowerCAmelCase = self.SP_CHAR_MAPPING.get(UpperCAmelCase__ ) else: lowerCAmelCase = unicodedata.normalize('NFKC' , UpperCAmelCase__ ) if self.is_whitespace(UpperCAmelCase__ ): continue normalized_text += ch char_mapping.extend([i] * len(UpperCAmelCase__ ) ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = normalized_text, [], 0 if self.do_lower_case: lowerCAmelCase = text.lower() for token in split_tokens: if token[:1] == "▁": lowerCAmelCase = token[1:] lowerCAmelCase = text[offset:].index(UpperCAmelCase__ ) + offset lowerCAmelCase = start + len(UpperCAmelCase__ ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowerCAmelCase = end return token_mapping @property def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: return len(self.vocab ) def __UpperCAmelCase ( self : Dict ) -> Optional[int]: return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self : Union[str, Any] ) -> Optional[Any]: lowerCAmelCase = self.__dict__.copy() lowerCAmelCase = None return state def __setstate__( self : int , UpperCAmelCase__ : Tuple ) -> Optional[Any]: lowerCAmelCase = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase = {} lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : List[str] ) -> Tuple: return "".join((self.SP_CHAR_MAPPING.get(UpperCAmelCase__ , UpperCAmelCase__ ) for c in text) ) def __UpperCAmelCase ( self : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Optional[int]=6_4 , UpperCAmelCase__ : Union[str, Any]=0.1 ) -> Dict: if self.sp_model_kwargs.get('enable_sampling' ) is True: lowerCAmelCase = True if self.sp_model_kwargs.get('alpha' ) is not None: lowerCAmelCase = self.sp_model_kwargs.get('alpha' ) if self.sp_model_kwargs.get('nbest_size' ) is not None: lowerCAmelCase = self.sp_model_kwargs.get('nbest_size' ) if not enable_sampling: lowerCAmelCase = self.sp_model.EncodeAsPieces(UpperCAmelCase__ ) else: lowerCAmelCase = self.sp_model.SampleEncodeAsPieces(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = [] for pi, piece in enumerate(UpperCAmelCase__ ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(UpperCAmelCase__ ) and pi != 0: new_pieces.append(UpperCAmelCase__ ) continue else: continue lowerCAmelCase = 0 for i, chunk in enumerate(UpperCAmelCase__ ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(UpperCAmelCase__ ) or self.is_punct(UpperCAmelCase__ ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(UpperCAmelCase__ ) lowerCAmelCase = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowerCAmelCase = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowerCAmelCase = i if len(UpperCAmelCase__ ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[str]: lowerCAmelCase = ''.join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , ' ' ).strip() return out_string def __UpperCAmelCase ( self : str , UpperCAmelCase__ : List[str] ) -> Any: lowerCAmelCase = self.convert_ids_to_tokens(UpperCAmelCase__ ) lowerCAmelCase = ''.join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , ' ' ).strip() return out_string def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : Union[str, Any] ) -> List[Any]: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def __UpperCAmelCase ( self : str , UpperCAmelCase__ : int ) -> Any: return self.reverse_vocab.get(UpperCAmelCase__ , self.unk_token ) def __UpperCAmelCase ( self : Tuple , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int]=None ) -> List[str]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Tuple=None ) -> List[str]: if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : int=False ) -> Any: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] + ([0] * len(UpperCAmelCase__ )) + [1] return [1] + ([0] * len(UpperCAmelCase__ )) + [1] def __UpperCAmelCase ( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: # called when `add_special_tokens` is True, so align with `build_inputs_with_special_tokens` method if token_ids_a is None: # [CLS] X [SEP] return (len(UpperCAmelCase__ ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(UpperCAmelCase__ ) + 1) + [1] * (len(UpperCAmelCase__ ) + 3) def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Tuple ) -> Optional[Any]: if "\u4e00" <= char <= "\u9fff": return True return False def __UpperCAmelCase ( self : Tuple , UpperCAmelCase__ : List[str] ) -> int: if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def __UpperCAmelCase ( self : Optional[Any] , UpperCAmelCase__ : List[str] ) -> Any: if char in ",;:.?!~,;:。?!《》【】": return True return False def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : Optional[int] ) -> List[str]: if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(UpperCAmelCase__ ) == 1: lowerCAmelCase = unicodedata.category(UpperCAmelCase__ ) if cat == "Zs": return True return False def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : Any ) -> Optional[Any]: lowerCAmelCase = {} with io.open(UpperCAmelCase__ , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(UpperCAmelCase__ ): lowerCAmelCase = line.rstrip('\n' ) lowerCAmelCase = int(UpperCAmelCase__ ) return token_to_idx def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: lowerCAmelCase = 0 if os.path.isdir(UpperCAmelCase__ ): lowerCAmelCase = os.path.join( UpperCAmelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: lowerCAmelCase = (filename_prefix + '-' if filename_prefix else '') + save_directory with open(UpperCAmelCase__ , 'w' , encoding='utf-8' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda UpperCAmelCase__ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ' Please check that the vocabulary is not corrupted!' ) lowerCAmelCase = token_index writer.write(token + '\n' ) index += 1 lowerCAmelCase = os.path.join(UpperCAmelCase__ , 'sentencepiece.bpe.model' ) with open(UpperCAmelCase__ , 'wb' ) as fi: lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (vocab_file,)
357
'''simple docstring''' import math def a_ ( lowerCamelCase : int ): lowerCAmelCase = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(lowerCamelCase ) def a_ ( lowerCamelCase : float = 1 / 12345 ): lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = 3 while True: lowerCAmelCase = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(lowerCamelCase ): lowerCAmelCase = int(lowerCamelCase ) total_partitions += 1 if check_partition_perfect(lowerCamelCase ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(lowerCamelCase ) integer += 1 if __name__ == "__main__": print(F'''{solution() = }''')
55
0
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : int = 0 __UpperCAmelCase : bool = False __UpperCAmelCase : float = 3.0 class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> List[str]: # If no defaults are changed, `to_kwargs` returns an empty dict. self.assertDictEqual(MockClass().to_kwargs() , {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs() , {"a": 2} ) self.assertDictEqual(MockClass(a=2 , b=lowerCamelCase ).to_kwargs() , {"a": 2, "b": True} ) self.assertDictEqual(MockClass(a=2 , c=2.25 ).to_kwargs() , {"a": 2, "c": 2.25} ) @require_cuda def __snake_case ( self : Optional[int] ) -> int: # If no defaults are changed, `to_kwargs` returns an empty dict. __snake_case : Any = GradScalerKwargs(init_scale=1024 , growth_factor=2 ) AcceleratorState._reset_state() __snake_case : Optional[Any] = Accelerator(mixed_precision="fp16" , kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) __snake_case : Optional[Any] = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale , 10_24.0 ) self.assertEqual(scaler._growth_factor , 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor , 0.5 ) self.assertEqual(scaler._growth_interval , 2000 ) self.assertEqual(scaler._enabled , lowerCamelCase ) @require_multi_gpu def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : List[str] = ["torchrun", F'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__ )] execute_subprocess_async(lowerCamelCase , env=os.environ.copy() ) if __name__ == "__main__": _snake_case : Optional[Any] = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) _snake_case : Optional[int] = Accelerator(kwargs_handlers=[ddp_scaler]) _snake_case : int = torch.nn.Linear(100, 200) _snake_case : str = accelerator.prepare(model) # Check the values changed in kwargs _snake_case : str = "" _snake_case : List[str] = model.bucket_bytes_cap // (1_024 * 1_024) if observed_bucket_cap_map != 15: error_msg += f"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += f"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += f"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += f"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += f"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
123
from importlib import import_module from .logging import get_logger _snake_case : Optional[int] = get_logger(__name__) class a : """simple docstring""" def __init__( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : List[str]=None ) -> Any: __snake_case : Dict = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("__" ): setattr(self , lowerCamelCase , getattr(lowerCamelCase , lowerCamelCase ) ) __snake_case : int = module._original_module if isinstance(lowerCamelCase , _PatchedModuleObj ) else module class a : """simple docstring""" __UpperCAmelCase : List[Any] = [] def __init__( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : str , lowerCamelCase : Dict , lowerCamelCase : Optional[Any]=None ) -> List[Any]: __snake_case : Union[str, Any] = obj __snake_case : Dict = target __snake_case : Any = new __snake_case : List[str] = target.split("." )[0] __snake_case : Union[str, Any] = {} __snake_case : int = attrs or [] def __enter__( self : List[Any] ) -> Tuple: *__snake_case , __snake_case : int = self.target.split("." ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowerCamelCase ) ): try: __snake_case : Any = import_module(".".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): __snake_case : Union[str, Any] = getattr(self.obj , lowerCamelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowerCamelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): __snake_case : List[Any] = obj_attr # patch at top level setattr(self.obj , lowerCamelCase , _PatchedModuleObj(lowerCamelCase , attrs=self.attrs ) ) __snake_case : Optional[int] = getattr(self.obj , lowerCamelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowerCamelCase , lowerCamelCase , _PatchedModuleObj(getattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) , attrs=self.attrs ) ) __snake_case : List[Any] = getattr(lowerCamelCase , lowerCamelCase ) # finally set the target attribute setattr(lowerCamelCase , lowerCamelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: __snake_case : Union[str, Any] = getattr(import_module(".".join(lowerCamelCase ) ) , lowerCamelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , lowerCamelCase ) is attr_value: __snake_case : Tuple = getattr(self.obj , lowerCamelCase ) setattr(self.obj , lowerCamelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" __snake_case : Dict = globals()["__builtins__"][target_attr] setattr(self.obj , lowerCamelCase , self.new ) else: raise RuntimeError(F'Tried to patch attribute {target_attr} instead of a submodule.' ) def __exit__( self : Any , *lowerCamelCase : Any ) -> Optional[int]: for attr in list(self.original ): setattr(self.obj , lowerCamelCase , self.original.pop(lowerCamelCase ) ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: self.__enter__() self._active_patches.append(self ) def __snake_case ( self : Any ) -> List[str]: try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
123
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class _snake_case : def __init__( self ,_snake_case ,_snake_case=13 ,_snake_case=7 ,_snake_case=True ,_snake_case=True ,_snake_case=True ,_snake_case=True ,_snake_case=99 ,_snake_case=32 ,_snake_case=2 ,_snake_case=4 ,_snake_case=37 ,_snake_case="gelu" ,_snake_case=0.1 ,_snake_case=0.1 ,_snake_case=5_12 ,_snake_case=16 ,_snake_case=2 ,_snake_case=0.02 ,_snake_case=3 ,_snake_case=4 ,_snake_case=None ,): UpperCAmelCase_ : List[Any] = parent UpperCAmelCase_ : str = 13 UpperCAmelCase_ : Optional[int] = 7 UpperCAmelCase_ : Union[str, Any] = True UpperCAmelCase_ : Union[str, Any] = True UpperCAmelCase_ : List[str] = True UpperCAmelCase_ : Union[str, Any] = True UpperCAmelCase_ : str = 99 UpperCAmelCase_ : int = 32 UpperCAmelCase_ : Dict = 2 UpperCAmelCase_ : Union[str, Any] = 4 UpperCAmelCase_ : Dict = 37 UpperCAmelCase_ : List[str] = """gelu""" UpperCAmelCase_ : List[str] = 0.1 UpperCAmelCase_ : Union[str, Any] = 0.1 UpperCAmelCase_ : List[Any] = 5_12 UpperCAmelCase_ : Optional[int] = 16 UpperCAmelCase_ : Dict = 2 UpperCAmelCase_ : List[str] = 0.02 UpperCAmelCase_ : List[Any] = 3 UpperCAmelCase_ : Any = 4 UpperCAmelCase_ : Union[str, Any] = None def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase_ : Union[str, Any] = None if self.use_input_mask: UpperCAmelCase_ : int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : Any = None if self.use_token_type_ids: UpperCAmelCase_ : Dict = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase_ : Any = None UpperCAmelCase_ : List[Any] = None UpperCAmelCase_ : Union[str, Any] = None if self.use_labels: UpperCAmelCase_ : List[Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase_ : Any = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase_ : int = RoFormerConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,return_dict=_lowerCAmelCase ,) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Optional[int] = TFRoFormerModel(config=_lowerCAmelCase ) UpperCAmelCase_ : List[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCAmelCase_ : Union[str, Any] = [input_ids, input_mask] UpperCAmelCase_ : Optional[int] = model(_lowerCAmelCase ) UpperCAmelCase_ : str = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : List[str] = True UpperCAmelCase_ : List[str] = TFRoFormerForCausalLM(config=_lowerCAmelCase ) UpperCAmelCase_ : List[Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCAmelCase_ : Union[str, Any] = model(_lowerCAmelCase )["""logits"""] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) ,[self.batch_size, self.seq_length, self.vocab_size] ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Any = TFRoFormerForMaskedLM(config=_lowerCAmelCase ) UpperCAmelCase_ : Union[str, Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCAmelCase_ : str = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : List[Any] = self.num_labels UpperCAmelCase_ : List[str] = TFRoFormerForSequenceClassification(config=_lowerCAmelCase ) UpperCAmelCase_ : Union[str, Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCAmelCase_ : Dict = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Dict = self.num_choices UpperCAmelCase_ : Optional[Any] = TFRoFormerForMultipleChoice(config=_lowerCAmelCase ) UpperCAmelCase_ : str = tf.tile(tf.expand_dims(_lowerCAmelCase ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase_ : Optional[int] = tf.tile(tf.expand_dims(_lowerCAmelCase ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase_ : List[str] = tf.tile(tf.expand_dims(_lowerCAmelCase ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase_ : List[str] = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } UpperCAmelCase_ : Tuple = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : str = self.num_labels UpperCAmelCase_ : List[Any] = TFRoFormerForTokenClassification(config=_lowerCAmelCase ) UpperCAmelCase_ : str = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCAmelCase_ : Optional[Any] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Optional[Any] = TFRoFormerForQuestionAnswering(config=_lowerCAmelCase ) UpperCAmelCase_ : Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCAmelCase_ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[Any] = self.prepare_config_and_inputs() ( UpperCAmelCase_ ) : Optional[int] = config_and_inputs UpperCAmelCase_ : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class _snake_case (__UpperCamelCase , __UpperCamelCase , unittest.TestCase): __A : Union[str, Any] =( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) __A : str =( { "feature-extraction": TFRoFormerModel, "fill-mask": TFRoFormerForMaskedLM, "question-answering": TFRoFormerForQuestionAnswering, "text-classification": TFRoFormerForSequenceClassification, "text-generation": TFRoFormerForCausalLM, "token-classification": TFRoFormerForTokenClassification, "zero-shot": TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) __A : Optional[int] =False __A : Optional[int] =False def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ): if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[Any] = TFRoFormerModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self ,config_class=_lowerCAmelCase ,hidden_size=37 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) @slow def UpperCamelCase__ ( self ): UpperCAmelCase_ : Union[str, Any] = TFRoFormerModel.from_pretrained("junnyu/roformer_chinese_base" ) self.assertIsNotNone(_lowerCAmelCase ) @require_tf class _snake_case (unittest.TestCase): @slow def UpperCamelCase__ ( self ): UpperCAmelCase_ : Any = TFRoFormerForMaskedLM.from_pretrained("junnyu/roformer_chinese_base" ) UpperCAmelCase_ : Union[str, Any] = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase_ : Dict = model(_lowerCAmelCase )[0] # TODO Replace vocab size UpperCAmelCase_ : Any = 5_00_00 UpperCAmelCase_ : Union[str, Any] = [1, 6, vocab_size] self.assertEqual(output.shape ,_lowerCAmelCase ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. UpperCAmelCase_ : str = tf.constant( [ [ [-0.12053341, -1.0264901, 0.29221946], [-1.5133783, 0.197433, 0.15190607], [-5.0135403, -3.900256, -0.84038764], ] ] ) tf.debugging.assert_near(output[:, :3, :3] ,_lowerCAmelCase ,atol=1E-4 ) @require_tf class _snake_case (unittest.TestCase): __A : List[str] =1e-4 def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[str] = tf.constant([[4, 10]] ) UpperCAmelCase_ : int = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 ,embedding_dim=6 ) UpperCAmelCase_ : Tuple = emba(input_ids.shape ) UpperCAmelCase_ : str = tf.constant( [[0.0000, 0.0000, 0.0000, 1.0000, 1.0000, 1.0000], [0.8415, 0.0464, 0.0022, 0.5403, 0.9989, 1.0000]] ) tf.debugging.assert_near(_lowerCAmelCase ,_lowerCAmelCase ,atol=self.tolerance ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[int] = tf.constant( [ [0.0000, 0.0000, 0.0000, 0.0000, 0.0000], [0.8415, 0.8219, 0.8020, 0.7819, 0.7617], [0.9093, 0.9364, 0.9581, 0.9749, 0.9870], ] ) UpperCAmelCase_ : List[Any] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=5_12 ,embedding_dim=5_12 ) emba([2, 16, 5_12] ) UpperCAmelCase_ : str = emba.weight[:3, :5] tf.debugging.assert_near(_lowerCAmelCase ,_lowerCAmelCase ,atol=self.tolerance ) @require_tf class _snake_case (unittest.TestCase): __A : Any =1e-4 def UpperCamelCase__ ( self ): # 2,12,16,64 UpperCAmelCase_ : Dict = tf.reshape(tf.range(2 * 12 * 16 * 64 ,dtype=tf.floataa ) ,shape=(2, 12, 16, 64) ) / 1_00 UpperCAmelCase_ : Optional[Any] = -tf.reshape(tf.range(2 * 12 * 16 * 64 ,dtype=tf.floataa ) ,shape=(2, 12, 16, 64) ) / 1_00 UpperCAmelCase_ : Union[str, Any] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32 ,embedding_dim=64 ) UpperCAmelCase_ : Any = embed_positions([2, 16, 7_68] )[None, None, :, :] UpperCAmelCase_ : Optional[int] = TFRoFormerSelfAttention.apply_rotary_position_embeddings( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) UpperCAmelCase_ : List[str] = tf.constant( [ [0.0000, 0.0100, 0.0200, 0.0300, 0.0400, 0.0500, 0.0600, 0.0700], [-0.2012, 0.8897, 0.0263, 0.9401, 0.2074, 0.9463, 0.3481, 0.9343], [-1.7057, 0.6271, -1.2145, 1.3897, -0.6303, 1.7647, -0.1173, 1.8985], [-2.1731, -1.6397, -2.7358, 0.2854, -2.1840, 1.7183, -1.3018, 2.4871], [0.2717, -3.6173, -2.9206, -2.1988, -3.6638, 0.3858, -2.9155, 2.2980], [3.9859, -2.1580, -0.7984, -4.4904, -4.1181, -2.0252, -4.4782, 1.1253], ] ) UpperCAmelCase_ : Tuple = tf.constant( [ [0.0000, -0.0100, -0.0200, -0.0300, -0.0400, -0.0500, -0.0600, -0.0700], [0.2012, -0.8897, -0.0263, -0.9401, -0.2074, -0.9463, -0.3481, -0.9343], [1.7057, -0.6271, 1.2145, -1.3897, 0.6303, -1.7647, 0.1173, -1.8985], [2.1731, 1.6397, 2.7358, -0.2854, 2.1840, -1.7183, 1.3018, -2.4871], [-0.2717, 3.6173, 2.9206, 2.1988, 3.6638, -0.3858, 2.9155, -2.2980], [-3.9859, 2.1580, 0.7984, 4.4904, 4.1181, 2.0252, 4.4782, -1.1253], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] ,_lowerCAmelCase ,atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] ,_lowerCAmelCase ,atol=self.tolerance )
370
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["""MobileViTFeatureExtractor"""] _lowerCamelCase = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class A_ ( unittest.TestCase ): """simple docstring""" def __init__( self :Union[str, Any] , lowerCamelCase_ :Union[str, Any] , lowerCamelCase_ :List[str]=7 , lowerCamelCase_ :Any=3 , lowerCamelCase_ :Dict=18 , lowerCamelCase_ :Union[str, Any]=30 , lowerCamelCase_ :List[Any]=400 , lowerCamelCase_ :Optional[int]=True , lowerCamelCase_ :Tuple=None , lowerCamelCase_ :Union[str, Any]=True , ): """simple docstring""" lowerCamelCase__ : Optional[int] =size if size is not None else {'height': 18, 'width': 18} lowerCamelCase__ : str =parent lowerCamelCase__ : Union[str, Any] =batch_size lowerCamelCase__ : Union[str, Any] =num_channels lowerCamelCase__ : int =image_size lowerCamelCase__ : Tuple =min_resolution lowerCamelCase__ : int =max_resolution lowerCamelCase__ : Tuple =do_resize lowerCamelCase__ : List[str] =size lowerCamelCase__ : str =do_normalize def UpperCAmelCase__ ( self :List[Any] ): """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.88_66_44_36_34_03_32_03, 0.66_18_82_93_69_54_49_83, 0.38_91_74_64_01_78_68_04], [-0.60_42_55_91_46_88_11_04, -0.0_22_95_00_88_60_52_84_69, 0.54_23_79_73_69_00_32_96], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class A_ ( A__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ImageGPTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self :Dict ): """simple docstring""" lowerCamelCase__ : int =ImageGPTImageProcessingTester(self ) @property def UpperCAmelCase__ ( self :Tuple ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self :Tuple ): """simple docstring""" lowerCamelCase__ : Any =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase_ , 'clusters' ) ) self.assertTrue(hasattr(lowerCamelCase_ , 'do_resize' ) ) self.assertTrue(hasattr(lowerCamelCase_ , 'size' ) ) self.assertTrue(hasattr(lowerCamelCase_ , 'do_normalize' ) ) def UpperCAmelCase__ ( self :Tuple ): """simple docstring""" lowerCamelCase__ : str =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 18, 'width': 18} ) lowerCamelCase__ : Tuple =self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) def UpperCAmelCase__ ( self :int ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =self.image_processing_class(**self.image_processor_dict ) lowerCamelCase__ : Tuple =json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase_ , obj[key] ) ) else: self.assertEqual(obj[key] , lowerCamelCase_ ) def UpperCAmelCase__ ( self :Tuple ): """simple docstring""" lowerCamelCase__ : int =self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : Optional[int] =os.path.join(lowerCamelCase_ , 'image_processor.json' ) image_processor_first.to_json_file(lowerCamelCase_ ) lowerCamelCase__ : List[Any] =self.image_processing_class.from_json_file(lowerCamelCase_ ).to_dict() lowerCamelCase__ : Optional[int] =image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowerCamelCase_ ) def UpperCAmelCase__ ( self :Optional[int] ): """simple docstring""" lowerCamelCase__ : List[Any] =self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(lowerCamelCase_ ) lowerCamelCase__ : str =self.image_processing_class.from_pretrained(lowerCamelCase_ ).to_dict() lowerCamelCase__ : Optional[Any] =image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowerCamelCase_ ) @unittest.skip('ImageGPT requires clusters at initialization' ) def UpperCAmelCase__ ( self :str ): """simple docstring""" pass def lowerCAmelCase_ ( ) ->Union[str, Any]: lowerCamelCase__ : Optional[Any] =load_dataset('hf-internal-testing/fixtures_image_utils' , split='test' ) lowerCamelCase__ : Optional[Any] =Image.open(dataset[4]['file'] ) lowerCamelCase__ : List[Any] =Image.open(dataset[5]['file'] ) lowerCamelCase__ : List[str] =[imagea, imagea] return images @require_vision @require_torch class A_ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self :Dict ): """simple docstring""" lowerCamelCase__ : Optional[int] =ImageGPTImageProcessor.from_pretrained('openai/imagegpt-small' ) lowerCamelCase__ : int =prepare_images() # test non-batched lowerCamelCase__ : Tuple =image_processing(images[0] , return_tensors='pt' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1_024) ) lowerCamelCase__ : Tuple =[306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , lowerCamelCase_ ) # test batched lowerCamelCase__ : List[str] =image_processing(lowerCamelCase_ , return_tensors='pt' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1_024) ) lowerCamelCase__ : int =[303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , lowerCamelCase_ )
126
"""simple docstring""" import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class A_ : """simple docstring""" @staticmethod def UpperCAmelCase__ ( *lowerCamelCase_ :Optional[int] , **lowerCamelCase_ :Optional[Any] ): """simple docstring""" pass def lowerCAmelCase_ ( snake_case_ : List[str] ) ->str: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. lowerCAmelCase = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class A_ ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Optional[Any] , lowerCamelCase_ :Union[str, Any] , lowerCamelCase_ :Tuple , lowerCamelCase_ :int ): """simple docstring""" lowerCamelCase__ : Optional[int] =pipeline( 'document-question-answering' , model=lowerCamelCase_ , tokenizer=lowerCamelCase_ , image_processor=lowerCamelCase_ ) lowerCamelCase__ : Tuple =INVOICE_URL lowerCamelCase__ : Optional[Any] =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) lowerCamelCase__ : Optional[Any] ='What is the placebo?' lowerCamelCase__ : List[str] =[ { 'image': load_image(lowerCamelCase_ ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def UpperCAmelCase__ ( self :int , lowerCamelCase_ :Tuple , lowerCamelCase_ :Tuple ): """simple docstring""" lowerCamelCase__ : List[str] =dqa_pipeline(lowerCamelCase_ , top_k=2 ) self.assertEqual( lowerCamelCase_ , [ [ {'score': ANY(lowerCamelCase_ ), 'answer': ANY(lowerCamelCase_ ), 'start': ANY(lowerCamelCase_ ), 'end': ANY(lowerCamelCase_ )}, {'score': ANY(lowerCamelCase_ ), 'answer': ANY(lowerCamelCase_ ), 'start': ANY(lowerCamelCase_ ), 'end': ANY(lowerCamelCase_ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : str =pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) lowerCamelCase__ : Any =INVOICE_URL lowerCamelCase__ : Union[str, Any] ='How many cats are there?' lowerCamelCase__ : List[Any] =[ {'score': 0.00_01, 'answer': 'oy 2312/2019', 'start': 38, 'end': 39}, {'score': 0.00_01, 'answer': 'oy 2312/2019 DUE', 'start': 38, 'end': 40}, ] lowerCamelCase__ : Dict =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , lowerCamelCase_ ) lowerCamelCase__ : int =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , lowerCamelCase_ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowerCamelCase__ : str ='./tests/fixtures/tests_samples/COCO/000000039769.png' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(lowerCamelCase_ , [] ) # We can optionnally pass directly the words and bounding boxes lowerCamelCase__ : str ='./tests/fixtures/tests_samples/COCO/000000039769.png' lowerCamelCase__ : Optional[Any] =[] lowerCamelCase__ : Tuple =[] lowerCamelCase__ : Tuple =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , words=lowerCamelCase_ , boxes=lowerCamelCase_ , top_k=2 ) self.assertEqual(lowerCamelCase_ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :Any ): """simple docstring""" lowerCamelCase__ : int =pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) lowerCamelCase__ : Dict =INVOICE_URL lowerCamelCase__ : int ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[str] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_44, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.00_09, 'answer': 'us-001', 'start': 16, 'end': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCAmelCase__ ( self :List[str] ): """simple docstring""" lowerCamelCase__ : int =pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=50 , ) lowerCamelCase__ : Tuple =INVOICE_URL lowerCamelCase__ : Any ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[Any] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : List[Any] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_74, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.99_48, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Dict ): """simple docstring""" lowerCamelCase__ : int =AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCamelCase_ ) lowerCamelCase__ : Optional[Any] =pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCamelCase_ , revision='3dc6de3' , ) lowerCamelCase__ : int =INVOICE_URL lowerCamelCase__ : Tuple ='What is the invoice number?' lowerCamelCase__ : Optional[Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowerCamelCase__ : Optional[Any] =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] ] * 2 , ) lowerCamelCase__ : Tuple =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) # This model should also work if `image` is set to None lowerCamelCase__ : Optional[int] =dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.42_51, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.08_19, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def UpperCAmelCase__ ( self :Optional[Any] ): """simple docstring""" lowerCamelCase__ : Optional[Any] =AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCamelCase_ ) lowerCamelCase__ : Any =pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCamelCase_ , revision='3dc6de3' , max_seq_len=50 , ) lowerCamelCase__ : Dict =INVOICE_URL lowerCamelCase__ : Optional[Any] ='What is the invoice number?' lowerCamelCase__ : Tuple =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowerCamelCase__ : Tuple =dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) lowerCamelCase__ : str =list(zip(*apply_tesseract(load_image(lowerCamelCase_ ) , lowerCamelCase_ , '' ) ) ) # This model should also work if `image` is set to None lowerCamelCase__ : Union[str, Any] =dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4 ) , [ {'score': 0.99_99, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.99_98, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) @slow @require_torch def UpperCAmelCase__ ( self :str ): """simple docstring""" lowerCamelCase__ : List[Any] =pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) lowerCamelCase__ : Union[str, Any] =INVOICE_URL lowerCamelCase__ : Union[str, Any] ='What is the invoice number?' lowerCamelCase__ : Union[str, Any] =dqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2 ) self.assertEqual(nested_simplify(lowerCamelCase_ , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def UpperCAmelCase__ ( self :str ): """simple docstring""" pass
126
1
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration UpperCAmelCase : int = { 'tiny.en': 'https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt', 'tiny': 'https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt', 'base.en': 'https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt', 'base': 'https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt', 'small.en': 'https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt', 'small': 'https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt', 'medium.en': 'https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt', 'medium': 'https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt', 'large': 'https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt', 'large-v2': 'https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt', } def _A ( SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" a__ : Any =["layers", "blocks"] for k in ignore_keys: state_dict.pop(UpperCamelCase__ , UpperCamelCase__ ) UpperCAmelCase : Any = { 'blocks': 'layers', 'mlp.0': 'fc1', 'mlp.2': 'fc2', 'mlp_ln': 'final_layer_norm', '.attn.query': '.self_attn.q_proj', '.attn.key': '.self_attn.k_proj', '.attn.value': '.self_attn.v_proj', '.attn_ln': '.self_attn_layer_norm', '.attn.out': '.self_attn.out_proj', '.cross_attn.query': '.encoder_attn.q_proj', '.cross_attn.key': '.encoder_attn.k_proj', '.cross_attn.value': '.encoder_attn.v_proj', '.cross_attn_ln': '.encoder_attn_layer_norm', '.cross_attn.out': '.encoder_attn.out_proj', 'decoder.ln.': 'decoder.layer_norm.', 'encoder.ln.': 'encoder.layer_norm.', 'token_embedding': 'embed_tokens', 'encoder.positional_embedding': 'encoder.embed_positions.weight', 'decoder.positional_embedding': 'decoder.embed_positions.weight', 'ln_post': 'layer_norm', } def _A ( SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" a__ : List[str] =list(s_dict.keys() ) for key in keys: a__ : str =key for k, v in WHISPER_MAPPING.items(): if k in key: a__ : List[str] =new_key.replace(UpperCamelCase__ , UpperCamelCase__ ) print(f'''{key} -> {new_key}''' ) a__ : Tuple =s_dict.pop(UpperCamelCase__ ) return s_dict def _A ( SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" a__ , a__ : Any =emb.weight.shape a__ : Any =nn.Linear(UpperCamelCase__ , UpperCamelCase__ , bias=UpperCamelCase__ ) a__ : List[Any] =emb.weight.data return lin_layer def _A ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str ): """simple docstring""" os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) a__ : Any =os.path.basename(UpperCamelCase__ ) a__ : List[Any] =url.split("/" )[-2] a__ : str =os.path.join(UpperCamelCase__ , UpperCamelCase__ ) if os.path.exists(UpperCamelCase__ ) and not os.path.isfile(UpperCamelCase__ ): raise RuntimeError(f'''{download_target} exists and is not a regular file''' ) if os.path.isfile(UpperCamelCase__ ): a__ : Dict =open(UpperCamelCase__ , "rb" ).read() if hashlib.shaaaa(UpperCamelCase__ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f'''{download_target} exists, but the SHA256 checksum does not match; re-downloading the file''' ) with urllib.request.urlopen(UpperCamelCase__ ) as source, open(UpperCamelCase__ , "wb" ) as output: with tqdm( total=int(source.info().get("Content-Length" ) ) , ncols=80 , unit="iB" , unit_scale=UpperCamelCase__ , unit_divisor=1_024 ) as loop: while True: a__ : str =source.read(8_192 ) if not buffer: break output.write(UpperCamelCase__ ) loop.update(len(UpperCamelCase__ ) ) a__ : int =open(UpperCamelCase__ , "rb" ).read() if hashlib.shaaaa(UpperCamelCase__ ).hexdigest() != expected_shaaaa: raise RuntimeError( "Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model." ) return model_bytes def _A ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" if ".pt" not in checkpoint_path: a__ : int =_download(_MODELS[checkpoint_path] ) else: a__ : List[str] =torch.load(UpperCamelCase__ , map_location="cpu" ) a__ : List[str] =original_checkpoint["dims"] a__ : List[Any] =original_checkpoint["model_state_dict"] a__ : Dict =state_dict["decoder.token_embedding.weight"] remove_ignore_keys_(UpperCamelCase__ ) rename_keys(UpperCamelCase__ ) a__ : List[Any] =True a__ : Dict =state_dict["decoder.layers.0.fc1.weight"].shape[0] a__ : Optional[int] =WhisperConfig( vocab_size=dimensions["n_vocab"] , encoder_ffn_dim=UpperCamelCase__ , decoder_ffn_dim=UpperCamelCase__ , num_mel_bins=dimensions["n_mels"] , d_model=dimensions["n_audio_state"] , max_target_positions=dimensions["n_text_ctx"] , encoder_layers=dimensions["n_audio_layer"] , encoder_attention_heads=dimensions["n_audio_head"] , decoder_layers=dimensions["n_text_layer"] , decoder_attention_heads=dimensions["n_text_state"] , max_source_positions=dimensions["n_audio_ctx"] , ) a__ : int =WhisperForConditionalGeneration(UpperCamelCase__ ) a__ , a__ : Tuple =model.model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0 and not set(UpperCamelCase__ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," f''' but all the following weights are missing {missing}''' ) if tie_embeds: a__ : Optional[Any] =make_linear_from_emb(model.model.decoder.embed_tokens ) else: a__ : Tuple =proj_out_weights model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": UpperCAmelCase : str = argparse.ArgumentParser() # # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Patht to the downloaded checkpoints""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") UpperCAmelCase : List[Any] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
367
UpperCAmelCase : Dict = [0, 2, 4, 6, 8] UpperCAmelCase : Tuple = [1, 3, 5, 7, 9] def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 a__ : str =0 for digit in range(10 ): a__ : int =digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return result a__ : List[str] =0 for digita in range(10 ): a__ : Optional[int] =digita if (remainder + digita) % 2 == 0: a__ : Dict =ODD_DIGITS else: a__ : Any =EVEN_DIGITS for digita in other_parity_digits: a__ : Union[str, Any] =digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) return result def _A ( SCREAMING_SNAKE_CASE : int = 9 ): """simple docstring""" a__ : List[str] =0 for length in range(1 , max_power + 1 ): result += reversible_numbers(SCREAMING_SNAKE_CASE , 0 , [0] * length , SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(F"""{solution() = }""")
148
0
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() __lowercase : Dict = logging.get_logger(__name__) def lowerCamelCase (_SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Any ): __a : List[str] = original_name.split('.' )[0] __a : Tuple = key.split('.' ) __a : Any = int(key_list[key_list.index(_SCREAMING_SNAKE_CASE ) - 2] ) __a : Union[str, Any] = int(key_list[key_list.index(_SCREAMING_SNAKE_CASE ) - 1] ) __a : List[Any] = orig_block_num - offset __a : List[Any] = key.replace(F"""{orig_block_num}.{layer_num}.{original_name}""" , F"""block.{new_block_num}.{layer_num}.{new_name}""" ) return key def lowerCamelCase (_SCREAMING_SNAKE_CASE : str ): __a : Optional[int] = OrderedDict() __a , __a : Union[str, Any] = 0, 0 for key, value in state_dict.items(): if key.startswith('network' ): __a : Optional[int] = key.replace('network' , 'poolformer.encoder' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('bias' ) and "patch_embed" not in key: patch_emb_offset += 1 __a : Dict = key[: key.find('proj' )] __a : Union[str, Any] = key.replace(_SCREAMING_SNAKE_CASE , F"""patch_embeddings.{total_embed_found}.""" ) __a : List[str] = key.replace('proj' , 'projection' ) if key.endswith('bias' ): total_embed_found += 1 if "patch_embeddings" in key: __a : str = 'poolformer.encoder.' + key if "mlp.fc1" in key: __a : Any = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'mlp.fc1' , 'output.conv1' ) if "mlp.fc2" in key: __a : str = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'mlp.fc2' , 'output.conv2' ) if "norm1" in key: __a : Tuple = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'norm1' , 'before_norm' ) if "norm2" in key: __a : str = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'norm2' , 'after_norm' ) if "layer_scale_1" in key: __a : Any = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'layer_scale_1' , 'layer_scale_1' ) if "layer_scale_2" in key: __a : Tuple = replace_key_with_offset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'layer_scale_2' , 'layer_scale_2' ) if "head" in key: __a : List[str] = key.replace('head' , 'classifier' ) __a : int = value return new_state_dict def lowerCamelCase (): __a : int = 'http://images.cocodataset.org/val2017/000000039769.jpg' __a : Optional[int] = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) return image @torch.no_grad() def lowerCamelCase (_SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Optional[Any] ): __a : Tuple = PoolFormerConfig() # set attributes based on model_name __a : str = 'huggingface/label-files' __a : str = model_name[-3:] __a : Optional[int] = 1_000 __a : Optional[int] = 'imagenet-1k-id2label.json' __a : Any = (1, 1_000) # set config attributes __a : Optional[int] = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) __a : Optional[Any] = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} __a : Optional[Any] = idalabel __a : Optional[Any] = {v: k for k, v in idalabel.items()} if size == "s12": __a : int = [2, 2, 6, 2] __a : str = [64, 128, 320, 512] __a : Tuple = 4.0 __a : List[Any] = 0.9 elif size == "s24": __a : Union[str, Any] = [4, 4, 12, 4] __a : str = [64, 128, 320, 512] __a : Optional[Any] = 4.0 __a : Tuple = 0.9 elif size == "s36": __a : str = [6, 6, 18, 6] __a : str = [64, 128, 320, 512] __a : str = 4.0 __a : Any = 1e-6 __a : int = 0.9 elif size == "m36": __a : Any = [6, 6, 18, 6] __a : str = [96, 192, 384, 768] __a : Dict = 4.0 __a : Optional[Any] = 1e-6 __a : List[str] = 0.9_5 elif size == "m48": __a : Union[str, Any] = [8, 8, 24, 8] __a : List[str] = [96, 192, 384, 768] __a : Tuple = 4.0 __a : List[Any] = 1e-6 __a : int = 0.9_5 else: raise ValueError(F"""Size {size} not supported""" ) # load image processor __a : List[Any] = PoolFormerImageProcessor(crop_pct=_SCREAMING_SNAKE_CASE ) # Prepare image __a : Dict = prepare_img() __a : Dict = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict __a : Optional[int] = torch.load(_SCREAMING_SNAKE_CASE , map_location=torch.device('cpu' ) ) # rename keys __a : Optional[int] = rename_keys(_SCREAMING_SNAKE_CASE ) # create HuggingFace model and load state dict __a : Any = PoolFormerForImageClassification(_SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) model.eval() # Define image processor __a : Tuple = PoolFormerImageProcessor(crop_pct=_SCREAMING_SNAKE_CASE ) __a : Tuple = image_processor(images=prepare_img() , return_tensors='pt' ).pixel_values # forward pass __a : int = model(_SCREAMING_SNAKE_CASE ) __a : Optional[int] = outputs.logits # define expected logit slices for different models if size == "s12": __a : str = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": __a : Tuple = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": __a : List[Any] = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": __a : Any = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": __a : Dict = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(F"""Size {size} not supported""" ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase : Tuple = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) __lowercase : Dict = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
27
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : Tuple ) -> Any: stooge(__lowerCamelCase , 0 , len(__lowerCamelCase ) - 1 ) return arr def _UpperCAmelCase ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : str ) -> int: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(__lowerCamelCase , __lowerCamelCase , (h - t) ) # Recursively sort last 2/3 elements stooge(__lowerCamelCase , i + t , (__lowerCamelCase) ) # Recursively sort first 2/3 elements stooge(__lowerCamelCase , __lowerCamelCase , (h - t) ) if __name__ == "__main__": UpperCAmelCase__ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase__ = [int(item) for item in user_input.split(',')] print(stooge_sort(unsorted))
288
0
"""simple docstring""" import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline __UpperCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) @dataclass class SCREAMING_SNAKE_CASE ( datasets.BuilderConfig ): """simple docstring""" lowercase__ = None lowercase__ = "utf-8" lowercase__ = None lowercase__ = None lowercase__ = True # deprecated lowercase__ = None # deprecated lowercase__ = 10 << 20 # 10MB lowercase__ = None class SCREAMING_SNAKE_CASE ( datasets.ArrowBasedBuilder ): """simple docstring""" lowercase__ = JsonConfig def __lowerCAmelCase ( self : Union[str, Any] ): if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowerCAmelCase__ : Optional[int] = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : Union[str, Any] ): if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}' ) lowerCAmelCase__ : Union[str, Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowercase_ ,(str, list, tuple) ): lowerCAmelCase__ : Optional[int] = data_files if isinstance(lowercase_ ,lowercase_ ): lowerCAmelCase__ : List[str] = [files] lowerCAmelCase__ : Union[str, Any] = [dl_manager.iter_files(lowercase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN ,gen_kwargs={'''files''': files} )] lowerCAmelCase__ : List[str] = [] for split_name, files in data_files.items(): if isinstance(lowercase_ ,lowercase_ ): lowerCAmelCase__ : Dict = [files] lowerCAmelCase__ : Optional[Any] = [dl_manager.iter_files(lowercase_ ) for file in files] splits.append(datasets.SplitGenerator(name=lowercase_ ,gen_kwargs={'''files''': files} ) ) return splits def __lowerCAmelCase ( self : int ,lowercase_ : pa.Table ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCAmelCase__ : int = self.config.features.arrow_schema.field(lowercase_ ).type lowerCAmelCase__ : List[Any] = pa_table.append_column(lowercase_ ,pa.array([None] * len(lowercase_ ) ,type=lowercase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase__ : Any = table_cast(lowercase_ ,self.config.features.arrow_schema ) return pa_table def __lowerCAmelCase ( self : List[str] ,lowercase_ : List[str] ): for file_idx, file in enumerate(itertools.chain.from_iterable(lowercase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(lowercase_ ,encoding=self.config.encoding ,errors=self.config.encoding_errors ) as f: lowerCAmelCase__ : Optional[int] = json.load(lowercase_ ) # We keep only the field we are interested in lowerCAmelCase__ : Dict = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(lowercase_ ,(list, tuple) ): lowerCAmelCase__ : str = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase__ : Any = {col: [row.get(lowercase_ ) for row in dataset] for col in keys} else: lowerCAmelCase__ : List[str] = dataset lowerCAmelCase__ : List[Any] = pa.Table.from_pydict(lowercase_ ) yield file_idx, self._cast_table(lowercase_ ) # If the file has one json object per line else: with open(lowercase_ ,'''rb''' ) as f: lowerCAmelCase__ : Dict = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCAmelCase__ : Optional[int] = max(self.config.chunksize // 3_2 ,1_6 << 1_0 ) lowerCAmelCase__ : List[Any] = ( self.config.encoding_errors if self.config.encoding_errors is not None else '''strict''' ) while True: lowerCAmelCase__ : List[str] = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(lowercase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCAmelCase__ : Optional[int] = batch.decode(self.config.encoding ,errors=lowercase_ ).encode('''utf-8''' ) try: while True: try: lowerCAmelCase__ : Optional[int] = paj.read_json( io.BytesIO(lowercase_ ) ,read_options=paj.ReadOptions(block_size=lowercase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(lowercase_ ,pa.ArrowInvalid ) and "straddling" not in str(lowercase_ ) or block_size > len(lowercase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F'Batch of {len(lowercase_ )} bytes couldn\'t be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.' ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( lowercase_ ,encoding=self.config.encoding ,errors=self.config.encoding_errors ) as f: lowerCAmelCase__ : Optional[Any] = json.load(lowercase_ ) except json.JSONDecodeError: logger.error(F'Failed to read file \'{file}\' with error {type(lowercase_ )}: {e}' ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(lowercase_ ,lowercase_ ): # list is the only sequence type supported in JSON try: lowerCAmelCase__ : Optional[int] = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase__ : str = {col: [row.get(lowercase_ ) for row in dataset] for col in keys} lowerCAmelCase__ : List[Any] = pa.Table.from_pydict(lowercase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F'Failed to read file \'{file}\' with error {type(lowercase_ )}: {e}' ) raise ValueError(F'Not able to read records in the JSON file at {file}.' ) from None yield file_idx, self._cast_table(lowercase_ ) break else: logger.error(F'Failed to read file \'{file}\' with error {type(lowercase_ )}: {e}' ) raise ValueError( F'Not able to read records in the JSON file at {file}. ' F'You should probably indicate the field of the JSON file containing your records. ' F'This JSON file contain the following fields: {str(list(dataset.keys() ) )}. ' F'Select the correct one and provide it as `field=\'XXX\'` to the dataset loading method. ' ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowercase_ ) batch_idx += 1
74
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): lowerCAmelCase__ : List[str] = Rectangle(height=0.5 ,width=0.5 ) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(6 )] lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[int] = VGroup(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''CPU''' ,font_size=2_4 ) lowerCAmelCase__ : Optional[int] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(4 )] lowerCAmelCase__ : List[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Any = Text('''GPU''' ,font_size=2_4 ) lowerCAmelCase__ : str = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[str] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Tuple = Text('''Model''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 / 4 ,width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ ,opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) ,buff=0.02 ,direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] ,direction=lowercase_ ,buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] ,direction=lowercase_ ,buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) lowerCAmelCase__ : List[str] = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[int] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''Loaded Checkpoint''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,aligned_edge=lowercase_ ,buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowerCAmelCase__ : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase__ : Optional[Any] = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' ,font_size=1_8 ,) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ ,lowercase_ ) lowerCAmelCase__ : Tuple = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' ,font_size=1_8 ,) blue_text.next_to(lowercase_ ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) lowerCAmelCase__ : Tuple = MarkupText( F'Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.' ,font_size=2_4 ,) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) ,Write(lowercase_ ) ) self.play(Write(lowercase_ ,run_time=1 ) ,Create(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : int = [] lowerCAmelCase__ : Optional[Any] = [] for i, rect in enumerate(lowercase_ ): lowerCAmelCase__ : List[str] = fill.copy().set_fill(lowercase_ ,opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : List[Any] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ ,run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
74
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json''', # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case_ ( __A ): __A : List[Any] = "blenderbot-small" __A : Tuple = ["past_key_values"] __A : Union[str, Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Any , lowercase_ : Any=5_02_65 , lowercase_ : Optional[Any]=5_12 , lowercase_ : Optional[int]=8 , lowercase_ : Tuple=20_48 , lowercase_ : Any=16 , lowercase_ : Optional[int]=8 , lowercase_ : Any=20_48 , lowercase_ : Any=16 , lowercase_ : Tuple=0.0 , lowercase_ : Optional[Any]=0.0 , lowercase_ : Union[str, Any]=True , lowercase_ : Optional[Any]=True , lowercase_ : int="gelu" , lowercase_ : str=5_12 , lowercase_ : str=0.1 , lowercase_ : Optional[int]=0.0 , lowercase_ : Union[str, Any]=0.0 , lowercase_ : Union[str, Any]=0.02 , lowercase_ : str=1 , lowercase_ : int=False , lowercase_ : Optional[int]=0 , lowercase_ : Tuple=1 , lowercase_ : int=2 , lowercase_ : List[str]=2 , **lowercase_ : Tuple , ) -> Union[str, Any]: lowercase__ : Any = vocab_size lowercase__ : int = max_position_embeddings lowercase__ : Optional[Any] = d_model lowercase__ : List[str] = encoder_ffn_dim lowercase__ : List[str] = encoder_layers lowercase__ : List[Any] = encoder_attention_heads lowercase__ : List[str] = decoder_ffn_dim lowercase__ : Optional[Any] = decoder_layers lowercase__ : Union[str, Any] = decoder_attention_heads lowercase__ : int = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Dict = activation_dropout lowercase__ : Union[str, Any] = activation_function lowercase__ : Dict = init_std lowercase__ : int = encoder_layerdrop lowercase__ : List[str] = decoder_layerdrop lowercase__ : str = use_cache lowercase__ : Dict = encoder_layers lowercase__ : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , decoder_start_token_id=lowercase_ , forced_eos_token_id=lowercase_ , **lowercase_ , ) class snake_case_ ( __A ): @property def __UpperCamelCase ( self : Any ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : str = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: lowercase__ : Tuple = {0: "batch"} lowercase__ : Any = {0: "batch", 1: "past_decoder_sequence + sequence"} else: lowercase__ : Dict = {0: "batch", 1: "decoder_sequence"} lowercase__ : Tuple = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowercase_ , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase__ : Optional[int] = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: lowercase__ , lowercase__ : Any = self.num_layers for i in range(lowercase_ ): lowercase__ : List[str] = {0: "batch", 2: "past_sequence + sequence"} lowercase__ : Any = {0: "batch", 2: "past_sequence + sequence"} else: lowercase__ : int = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def __UpperCamelCase ( self : str ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : Dict = super().outputs else: lowercase__ : List[str] = super(lowercase_ , self ).outputs if self.use_past: lowercase__ , lowercase__ : Optional[Any] = self.num_layers for i in range(lowercase_ ): lowercase__ : Dict = {0: "batch", 2: "past_sequence + sequence"} lowercase__ : List[Any] = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def __UpperCamelCase ( self : Tuple , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowercase__ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Generate decoder inputs lowercase__ : str = seq_length if not self.use_past else 1 lowercase__ : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase__ : Union[str, Any] = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} lowercase__ : Union[str, Any] = dict(**lowercase_ , **lowercase_ ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowercase__ , lowercase__ : Union[str, Any] = common_inputs["input_ids"].shape lowercase__ : Optional[int] = common_inputs["decoder_input_ids"].shape[1] lowercase__ , lowercase__ : List[str] = self.num_attention_heads lowercase__ : Dict = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ : List[str] = decoder_seq_length + 3 lowercase__ : Union[str, Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase__ : Tuple = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowercase_ , lowercase_ )] , dim=1 ) lowercase__ : Union[str, Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase__ , lowercase__ : List[str] = self.num_layers lowercase__ : List[Any] = min(lowercase_ , lowercase_ ) lowercase__ : List[Any] = max(lowercase_ , lowercase_ ) - min_num_layers lowercase__ : int = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowercase_ ): common_inputs["past_key_values"].append( ( torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), ) ) # TODO: test this. lowercase__ : str = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowercase_ , lowercase_ ): common_inputs["past_key_values"].append((torch.zeros(lowercase_ ), torch.zeros(lowercase_ )) ) return common_inputs def __UpperCamelCase ( self : Optional[Any] , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowercase__ : Any = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowercase__ , lowercase__ : str = common_inputs["input_ids"].shape # Not using the same length for past_key_values lowercase__ : Dict = seqlen + 2 lowercase__ , lowercase__ : List[str] = self.num_layers lowercase__ , lowercase__ : Optional[Any] = self.num_attention_heads lowercase__ : Optional[int] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ : Optional[int] = common_inputs["attention_mask"].dtype lowercase__ : List[Any] = torch.cat( [common_inputs["attention_mask"], torch.ones(lowercase_ , lowercase_ , dtype=lowercase_ )] , dim=1 ) lowercase__ : Dict = [ (torch.zeros(lowercase_ ), torch.zeros(lowercase_ )) for _ in range(lowercase_ ) ] return common_inputs def __UpperCamelCase ( self : List[Any] , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase__ : List[Any] = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase__ : Optional[Any] = tokenizer.num_special_tokens_to_add(lowercase_ ) lowercase__ : List[Any] = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowercase_ ) # Generate dummy inputs according to compute batch and sequence lowercase__ : int = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase__ : Union[str, Any] = dict(tokenizer(lowercase_ , return_tensors=lowercase_ ) ) return common_inputs def __UpperCamelCase ( self : str , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : str = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowercase_ , batch_size=lowercase_ , seq_length=lowercase_ , is_pair=lowercase_ , framework=lowercase_ ) elif self.task == "causal-lm": lowercase__ : List[str] = self._generate_dummy_inputs_for_causal_lm( lowercase_ , batch_size=lowercase_ , seq_length=lowercase_ , is_pair=lowercase_ , framework=lowercase_ ) else: lowercase__ : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowercase_ , batch_size=lowercase_ , seq_length=lowercase_ , is_pair=lowercase_ , framework=lowercase_ ) return common_inputs def __UpperCamelCase ( self : Tuple , lowercase_ : Tuple , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : List[Any] ) -> Any: if self.task in ["default", "seq2seq-lm"]: lowercase__ : Dict = super()._flatten_past_key_values_(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: lowercase__ : str = super(lowercase_ , self )._flatten_past_key_values_( lowercase_ , lowercase_ , lowercase_ , lowercase_ )
87
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. a : int = abspath(join(dirname(__file__), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' config.addinivalue_line( '''markers''', '''is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested''' ) config.addinivalue_line( '''markers''', '''is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested''' ) config.addinivalue_line('''markers''', '''is_pipeline_test: mark test to run only when pipelines are tested''' ) config.addinivalue_line('''markers''', '''is_staging_test: mark test to run only in the staging environment''' ) config.addinivalue_line('''markers''', '''accelerate_tests: mark test that require accelerate''' ) config.addinivalue_line('''markers''', '''tool_tests: mark the tool tests that are run on their specific schedule''' ) def __magic_name__ ( __UpperCAmelCase ) -> int: '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main snake_case_ = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(__UpperCAmelCase, id=__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if exitstatus == 5: snake_case_ = 0 # Doctest custom flag to ignore output. a : Union[str, Any] = doctest.register_optionflag('IGNORE_RESULT') a : Optional[int] = doctest.OutputChecker class a ( _lowerCamelCase ): def A_ ( self : List[Any] , lowercase_ : int , lowercase_ : Tuple , lowercase_ : Optional[int] ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , lowercase_ , lowercase_ , lowercase_ ) a : List[Any] = CustomOutputChecker a : Optional[int] = HfDoctestModule a : Tuple = HfDocTestParser
56
0
'''simple docstring''' from __future__ import annotations def a__ ( lowercase : int, lowercase : str, lowercase : Dict, lowercase : str ) -> Union[str, Any]: # noqa: E741 """simple docstring""" while r - l > 1: _UpperCamelCase = (l + r) // 2 if v[m] >= key: _UpperCamelCase = m else: _UpperCamelCase = m # noqa: E741 return r def a__ ( lowercase : list[int] ) -> int: """simple docstring""" if len(lowercase ) == 0: return 0 _UpperCamelCase = [0] * len(lowercase ) _UpperCamelCase = 1 _UpperCamelCase = v[0] for i in range(1, len(lowercase ) ): if v[i] < tail[0]: _UpperCamelCase = v[i] elif v[i] > tail[length - 1]: _UpperCamelCase = v[i] length += 1 else: _UpperCamelCase = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
287
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process lowercase__ : Any = logging.getLogger(__name__) def a__ ( lowercase : Optional[Any], lowercase : Tuple ) -> Any: """simple docstring""" return (preds == labels).mean() @dataclass class __lowerCAmelCase : """simple docstring""" _snake_case : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) _snake_case : Optional[str] = field( default=__magic_name__ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) _snake_case : Optional[str] = field( default=__magic_name__ , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) _snake_case : Optional[str] = field( default=__magic_name__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class __lowerCAmelCase : """simple docstring""" _snake_case : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) _snake_case : str = field(metadata={'help': 'Should contain the data files for the task.'} ) _snake_case : int = field( default=1_2_8 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) _snake_case : bool = field( default=__magic_name__ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a__ ( ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN, ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''', training_args.local_rank, training_args.device, training_args.n_gpu, bool(training_args.local_rank != -1 ), training_args.fpaa, ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''', lowercase ) # Set seed set_seed(training_args.seed ) try: _UpperCamelCase = processors[data_args.task_name]() _UpperCamelCase = processor.get_labels() _UpperCamelCase = len(lowercase ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCamelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=lowercase, finetuning_task=data_args.task_name, cache_dir=model_args.cache_dir, ) _UpperCamelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, cache_dir=model_args.cache_dir, ) _UpperCamelCase = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path, from_tf=bool('''.ckpt''' in model_args.model_name_or_path ), config=lowercase, cache_dir=model_args.cache_dir, ) # Get datasets _UpperCamelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir, tokenizer=lowercase, task=data_args.task_name, max_seq_length=data_args.max_seq_length, overwrite_cache=data_args.overwrite_cache, mode=Split.train, ) if training_args.do_train else None ) _UpperCamelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir, tokenizer=lowercase, task=data_args.task_name, max_seq_length=data_args.max_seq_length, overwrite_cache=data_args.overwrite_cache, mode=Split.dev, ) if training_args.do_eval else None ) def compute_metrics(lowercase : EvalPrediction ) -> Dict: _UpperCamelCase = np.argmax(p.predictions, axis=1 ) return {"acc": simple_accuracy(lowercase, p.label_ids )} # Data collator _UpperCamelCase = DataCollatorWithPadding(lowercase, pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer _UpperCamelCase = Trainer( model=lowercase, args=lowercase, train_dataset=lowercase, eval_dataset=lowercase, compute_metrics=lowercase, data_collator=lowercase, ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation _UpperCamelCase = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) _UpperCamelCase = trainer.evaluate() _UpperCamelCase = os.path.join(training_args.output_dir, '''eval_results.txt''' ) if trainer.is_world_master(): with open(lowercase, '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''', lowercase, lowercase ) writer.write('''%s = %s\n''' % (key, value) ) results.update(lowercase ) return results def a__ ( lowercase : Tuple ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
287
1
'''simple docstring''' from __future__ import annotations def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): A_, A_ : Dict = array[indexa], array[indexa] def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if length > 1: A_ : Dict = int(length / 2 ) for i in range(lowerCamelCase__ , low + middle ): comp_and_swap(lowerCamelCase__ , lowerCamelCase__ , i + middle , lowerCamelCase__ ) bitonic_merge(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) bitonic_merge(lowerCamelCase__ , low + middle , lowerCamelCase__ , lowerCamelCase__ ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if length > 1: A_ : List[str] = int(length / 2 ) bitonic_sort(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , 1 ) bitonic_sort(lowerCamelCase__ , low + middle , lowerCamelCase__ , 0 ) bitonic_merge(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if __name__ == "__main__": lowerCamelCase :Union[str, Any] = input('''Enter numbers separated by a comma:\n''').strip() lowerCamelCase :List[str] = [int(item.strip()) for item in user_input.split(''',''')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('''\nSorted array in ascending order is: ''', end='''''') print(*unsorted, sep=''', ''') bitonic_merge(unsorted, 0, len(unsorted), 0) print('''Sorted array in descending order is: ''', end='''''') print(*unsorted, sep=''', ''')
206
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Optional[int] = VideoMAEConfig() set_architecture_configs(lowerCamelCase__ , lowerCamelCase__ ) if "finetuned" not in model_name: A_ : Dict = False if "finetuned" in model_name: A_ : List[Any] = """huggingface/label-files""" if "kinetics" in model_name: A_ : Dict = 4_00 A_ : List[str] = """kinetics400-id2label.json""" elif "ssv2" in model_name: A_ : Tuple = 1_74 A_ : str = """something-something-v2-id2label.json""" else: raise ValueError("""Model name should either contain 'kinetics' or 'ssv2' in case it's fine-tuned.""" ) A_ : Dict = json.load(open(hf_hub_download(lowerCamelCase__ , lowerCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) A_ : List[str] = {int(lowerCamelCase__ ): v for k, v in idalabel.items()} A_ : Optional[Any] = idalabel A_ : Union[str, Any] = {v: k for k, v in idalabel.items()} return config def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if "small" in model_name: A_ : int = 3_84 A_ : Union[str, Any] = 15_36 A_ : List[str] = 12 A_ : Optional[int] = 16 A_ : Any = 12 A_ : int = 3 A_ : Optional[Any] = 1_92 A_ : Union[str, Any] = 7_68 elif "large" in model_name: A_ : List[Any] = 10_24 A_ : Optional[Any] = 40_96 A_ : Optional[Any] = 24 A_ : List[str] = 16 A_ : Any = 12 A_ : str = 8 A_ : str = 5_12 A_ : int = 20_48 elif "huge" in model_name: A_ : Optional[Any] = 12_80 A_ : str = 51_20 A_ : str = 32 A_ : int = 16 A_ : Any = 12 A_ : Union[str, Any] = 8 A_ : Dict = 6_40 A_ : Optional[Any] = 25_60 elif "base" not in model_name: raise ValueError("""Model name should include either \"small\", \"base\", \"large\", or \"huge\"""" ) def a ( lowerCamelCase__ ): '''simple docstring''' if "encoder." in name: A_ : List[Any] = name.replace("""encoder.""" , """""" ) if "cls_token" in name: A_ : List[str] = name.replace("""cls_token""" , """videomae.embeddings.cls_token""" ) if "decoder_pos_embed" in name: A_ : Tuple = name.replace("""decoder_pos_embed""" , """decoder.decoder_pos_embed""" ) if "pos_embed" in name and "decoder" not in name: A_ : int = name.replace("""pos_embed""" , """videomae.embeddings.position_embeddings""" ) if "patch_embed.proj" in name: A_ : Optional[Any] = name.replace("""patch_embed.proj""" , """videomae.embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: A_ : Dict = name.replace("""patch_embed.norm""" , """videomae.embeddings.norm""" ) if "decoder.blocks" in name: A_ : List[str] = name.replace("""decoder.blocks""" , """decoder.decoder_layers""" ) if "blocks" in name: A_ : List[str] = name.replace("""blocks""" , """videomae.encoder.layer""" ) if "attn.proj" in name: A_ : str = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name and "bias" not in name: A_ : str = name.replace("""attn""" , """attention.self""" ) if "attn" in name: A_ : Union[str, Any] = name.replace("""attn""" , """attention.attention""" ) if "norm1" in name: A_ : Any = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: A_ : List[str] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: A_ : Dict = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: A_ : List[str] = name.replace("""mlp.fc2""" , """output.dense""" ) if "decoder_embed" in name: A_ : Optional[Any] = name.replace("""decoder_embed""" , """decoder.decoder_embed""" ) if "decoder_norm" in name: A_ : Tuple = name.replace("""decoder_norm""" , """decoder.decoder_norm""" ) if "decoder_pred" in name: A_ : Tuple = name.replace("""decoder_pred""" , """decoder.decoder_pred""" ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: A_ : Dict = name.replace("""norm.weight""" , """videomae.layernorm.weight""" ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: A_ : List[str] = name.replace("""norm.bias""" , """videomae.layernorm.bias""" ) if "head" in name and "decoder" not in name: A_ : Optional[Any] = name.replace("""head""" , """classifier""" ) return name def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' for key in orig_state_dict.copy().keys(): A_ : str = orig_state_dict.pop(lowerCamelCase__ ) if key.startswith("""encoder.""" ): A_ : Tuple = key.replace("""encoder.""" , """""" ) if "qkv" in key: A_ : Optional[int] = key.split(""".""" ) if key.startswith("""decoder.blocks""" ): A_ : Union[str, Any] = config.decoder_hidden_size A_ : Any = int(key_split[2] ) A_ : int = """decoder.decoder_layers.""" if "weight" in key: A_ : Optional[Any] = val[:dim, :] A_ : Any = val[dim : dim * 2, :] A_ : Dict = val[-dim:, :] else: A_ : List[Any] = config.hidden_size A_ : List[Any] = int(key_split[1] ) A_ : int = """videomae.encoder.layer.""" if "weight" in key: A_ : Any = val[:dim, :] A_ : Union[str, Any] = val[dim : dim * 2, :] A_ : List[str] = val[-dim:, :] else: A_ : Union[str, Any] = val return orig_state_dict def a ( ): '''simple docstring''' A_ : List[Any] = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""" , filename="""eating_spaghetti.npy""" , repo_type="""dataset""" ) A_ : Optional[Any] = np.load(lowerCamelCase__ ) return list(lowerCamelCase__ ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = get_videomae_config(lowerCamelCase__ ) if "finetuned" in model_name: A_ : List[str] = VideoMAEForVideoClassification(lowerCamelCase__ ) else: A_ : Optional[Any] = VideoMAEForPreTraining(lowerCamelCase__ ) # download original checkpoint, hosted on Google Drive A_ : Optional[Any] = """pytorch_model.bin""" gdown.cached_download(lowerCamelCase__ , lowerCamelCase__ , quiet=lowerCamelCase__ ) A_ : Any = torch.load(lowerCamelCase__ , map_location="""cpu""" ) if "model" in files: A_ : Any = files["""model"""] else: A_ : Dict = files["""module"""] A_ : Any = convert_state_dict(lowerCamelCase__ , lowerCamelCase__ ) model.load_state_dict(lowerCamelCase__ ) model.eval() # verify model on basic input A_ : int = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) A_ : Union[str, Any] = prepare_video() A_ : str = image_processor(lowerCamelCase__ , return_tensors="""pt""" ) if "finetuned" not in model_name: A_ : List[str] = hf_hub_download(repo_id="""hf-internal-testing/bool-masked-pos""" , filename="""bool_masked_pos.pt""" ) A_ : Optional[Any] = torch.load(lowerCamelCase__ ) A_ : Dict = model(**lowerCamelCase__ ) A_ : List[Any] = outputs.logits A_ : Any = [ """videomae-small-finetuned-kinetics""", """videomae-small-finetuned-ssv2""", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) """videomae-base-short""", """videomae-base-short-finetuned-kinetics""", """videomae-base""", """videomae-base-finetuned-kinetics""", """videomae-large""", """videomae-large-finetuned-kinetics""", """videomae-huge-finetuned-kinetics""", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) """videomae-base-short-ssv2""", """videomae-base-short-finetuned-ssv2""", """videomae-base-ssv2""", """videomae-base-finetuned-ssv2""", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": A_ : str = torch.Size([1, 4_00] ) A_ : Optional[Any] = torch.tensor([-0.9_291, -0.4_061, -0.9_307] ) elif model_name == "videomae-small-finetuned-ssv2": A_ : str = torch.Size([1, 1_74] ) A_ : Union[str, Any] = torch.tensor([0.2_671, -0.4_689, -0.8_235] ) elif model_name == "videomae-base": A_ : Tuple = torch.Size([1, 14_08, 15_36] ) A_ : List[str] = torch.tensor([[0.7_739, 0.7_968, 0.7_089], [0.6_701, 0.7_487, 0.6_209], [0.4_287, 0.5_158, 0.4_773]] ) elif model_name == "videomae-base-short": A_ : Dict = torch.Size([1, 14_08, 15_36] ) A_ : List[str] = torch.tensor([[0.7_994, 0.9_612, 0.8_508], [0.7_401, 0.8_958, 0.8_302], [0.5_862, 0.7_468, 0.7_325]] ) # we verified the loss both for normalized and unnormalized targets for this one A_ : List[Any] = torch.tensor([0.5_142] ) if config.norm_pix_loss else torch.tensor([0.6_469] ) elif model_name == "videomae-large": A_ : str = torch.Size([1, 14_08, 15_36] ) A_ : Dict = torch.tensor([[0.7_149, 0.7_997, 0.6_966], [0.6_768, 0.7_869, 0.6_948], [0.5_139, 0.6_221, 0.5_605]] ) elif model_name == "videomae-large-finetuned-kinetics": A_ : int = torch.Size([1, 4_00] ) A_ : Optional[Any] = torch.tensor([0.0_771, 0.0_011, -0.3_625] ) elif model_name == "videomae-huge-finetuned-kinetics": A_ : Union[str, Any] = torch.Size([1, 4_00] ) A_ : Optional[int] = torch.tensor([0.2_433, 0.1_632, -0.4_894] ) elif model_name == "videomae-base-short-finetuned-kinetics": A_ : List[Any] = torch.Size([1, 4_00] ) A_ : Optional[Any] = torch.tensor([0.6_588, 0.0_990, -0.2_493] ) elif model_name == "videomae-base-finetuned-kinetics": A_ : Union[str, Any] = torch.Size([1, 4_00] ) A_ : Tuple = torch.tensor([0.3_669, -0.0_688, -0.2_421] ) elif model_name == "videomae-base-short-ssv2": A_ : Optional[Any] = torch.Size([1, 14_08, 15_36] ) A_ : List[Any] = torch.tensor([[0.4_712, 0.5_296, 0.5_786], [0.2_278, 0.2_729, 0.4_026], [0.0_352, 0.0_730, 0.2_506]] ) elif model_name == "videomae-base-short-finetuned-ssv2": A_ : Any = torch.Size([1, 1_74] ) A_ : Any = torch.tensor([-0.0_537, -0.1_539, -0.3_266] ) elif model_name == "videomae-base-ssv2": A_ : Dict = torch.Size([1, 14_08, 15_36] ) A_ : Dict = torch.tensor([[0.8_131, 0.8_727, 0.8_546], [0.7_366, 0.9_377, 0.8_870], [0.5_935, 0.8_874, 0.8_564]] ) elif model_name == "videomae-base-finetuned-ssv2": A_ : Any = torch.Size([1, 1_74] ) A_ : str = torch.tensor([0.1_961, -0.8_337, -0.6_389] ) else: raise ValueError(f'Model name not supported. Should be one of {model_names}' ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , lowerCamelCase__ , atol=1E-4 ) else: print("""Logits:""" , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , lowerCamelCase__ , atol=1E-4 ) print("""Logits ok!""" ) # verify loss, if applicable if model_name == "videomae-base-short": A_ : Optional[int] = outputs.loss assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-4 ) print("""Loss ok!""" ) if pytorch_dump_folder_path is not None: print(f'Saving model and image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(lowerCamelCase__ ) model.save_pretrained(lowerCamelCase__ ) if push_to_hub: print("""Pushing to the hub...""" ) model.push_to_hub(lowerCamelCase__ , organization="""nielsr""" ) if __name__ == "__main__": lowerCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4''', type=str, help=( '''URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct''' ''' download link.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''/Users/nielsrogge/Documents/VideoMAE/Test''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--model_name''', default='''videomae-base''', type=str, help='''Name of the model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCamelCase :Union[str, Any] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
206
1
def __snake_case ( _UpperCAmelCase ): __a = [] for data in source_data: for i, el in enumerate(snake_case__ ): if len(snake_case__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(snake_case__ ) ) return data_lists def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = [] for dlist, weight in zip(snake_case__ , snake_case__ ): __a = min(snake_case__ ) __a = max(snake_case__ ) __a = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: __a = f'Invalid weight of {weight:f} provided' raise ValueError(snake_case__ ) score_lists.append(snake_case__ ) return score_lists def __snake_case ( _UpperCAmelCase ): __a = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(snake_case__ ): __a = final_scores[j] + ele return final_scores def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = get_data(snake_case__ ) __a = calculate_each_score(snake_case__ , snake_case__ ) __a = generate_final_scores(snake_case__ ) # append scores to source data for i, ele in enumerate(snake_case__ ): source_data[i].append(snake_case__ ) return source_data
365
# Lint as: python3 import itertools import os import re __snake_case :int = re.compile(r'''([A-Z]+)([A-Z][a-z])''') __snake_case :List[Any] = re.compile(r'''([a-z\d])([A-Z])''') __snake_case :Dict = re.compile(r'''(?<!_)_(?!_)''') __snake_case :Optional[int] = re.compile(r'''(_{2,})''') __snake_case :Optional[Any] = r'''^\w+(\.\w+)*$''' __snake_case :Union[str, Any] = r'''<>:/\|?*''' def __snake_case ( _UpperCAmelCase ): __a = _uppercase_uppercase_re.sub(R'''\1_\2''' , _UpperCAmelCase ) __a = _lowercase_uppercase_re.sub(R'''\1_\2''' , _UpperCAmelCase ) return name.lower() def __snake_case ( _UpperCAmelCase ): __a = _single_underscore_re.split(_UpperCAmelCase ) __a = [_multiple_underscores_re.split(_UpperCAmelCase ) for n in name] return "".join(n.capitalize() for n in itertools.chain.from_iterable(_UpperCAmelCase ) if n != '''''' ) def __snake_case ( _UpperCAmelCase ): if os.path.basename(_UpperCAmelCase ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) return camelcase_to_snakecase(_UpperCAmelCase ) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): if os.path.basename(_UpperCAmelCase ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) if not re.match(_split_re , _UpperCAmelCase ): raise ValueError(f'Split name should match \'{_split_re}\'\' but got \'{split}\'.' ) return f'{filename_prefix_for_name(_UpperCAmelCase )}-{split}' def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None ): __a = filename_prefix_for_split(_UpperCAmelCase , _UpperCAmelCase ) if filetype_suffix: prefix += f'.{filetype_suffix}' __a = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) return f'{filepath}*' def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None ): __a = filename_prefix_for_split(_UpperCAmelCase , _UpperCAmelCase ) __a = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if shard_lengths: __a = len(_UpperCAmelCase ) __a = [f'{prefix}-{shard_id:05d}-of-{num_shards:05d}' for shard_id in range(_UpperCAmelCase )] if filetype_suffix: __a = [filename + f'.{filetype_suffix}' for filename in filenames] return filenames else: __a = prefix if filetype_suffix: filename += f'.{filetype_suffix}' return [filename]
131
0
def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Tuple: '''simple docstring''' UpperCAmelCase = [[0 for _ in range(UpperCAmelCase_ )] for _ in range(m + 1 )] for i in range(m + 1 ): UpperCAmelCase = 1 for n in range(m + 1 ): for k in range(1 , UpperCAmelCase_ ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: __A : Optional[int] = int(input("Enter a number: ").strip()) print(partition(n)) except ValueError: print("Please enter a number.") else: try: __A : str = int(sys.argv[1]) print(partition(n)) except ValueError: print("Please pass a number.")
273
'''simple docstring''' import math def __snake_case ( UpperCAmelCase_ : int ): lowerCamelCase_ = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(UpperCAmelCase_ ) def __snake_case ( UpperCAmelCase_ : float = 1 / 12345 ): lowerCamelCase_ = 0 lowerCamelCase_ = 0 lowerCamelCase_ = 3 while True: lowerCamelCase_ = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(UpperCAmelCase_ ): lowerCamelCase_ = int(UpperCAmelCase_ ) total_partitions += 1 if check_partition_perfect(UpperCAmelCase_ ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(UpperCAmelCase_ ) integer += 1 if __name__ == "__main__": print(f'''{solution() = }''')
55
0
"""simple docstring""" import math def UpperCamelCase ( UpperCAmelCase ) ->bool: """simple docstring""" a_ = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase = 1 / 12_345 ) ->int: """simple docstring""" a_ = 0 a_ = 0 a_ = 3 while True: a_ = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(UpperCAmelCase ): a_ = int(UpperCAmelCase ) total_partitions += 1 if check_partition_perfect(UpperCAmelCase ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(UpperCAmelCase ) integer += 1 if __name__ == "__main__": print(F"""{solution() = }""")
303
"""simple docstring""" import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class snake_case : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=32 , __UpperCAmelCase=5 , __UpperCAmelCase=4 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=50 , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , __UpperCAmelCase=None , ) ->Dict: a_ = parent a_ = batch_size a_ = seq_length a_ = is_training a_ = use_input_mask a_ = vocab_size a_ = hidden_size a_ = num_hidden_layers a_ = num_attention_heads a_ = intermediate_size a_ = hidden_act a_ = hidden_dropout_prob a_ = attention_probs_dropout_prob a_ = max_position_embeddings a_ = initializer_range a_ = use_labels a_ = scope def UpperCAmelCase__ ( self) ->Any: a_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a_ = None if self.use_input_mask: a_ = random_attention_mask([self.batch_size, self.seq_length]) if self.use_labels: a_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a_ = self.get_config() return config, input_ids, input_mask, token_labels def UpperCAmelCase__ ( self) ->Optional[Any]: return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self) ->List[str]: ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) = self.prepare_config_and_inputs() a_ = True a_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) a_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) ->str: a_ = BertGenerationEncoder(config=__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase) a_ = model(__UpperCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) ->Union[str, Any]: a_ = True a_ = BertGenerationEncoder(config=__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , ) a_ = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) ->List[str]: a_ = True a_ = True a_ = BertGenerationDecoder(config=__UpperCAmelCase).to(__UpperCAmelCase).eval() # first forward pass a_ = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase , ) a_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids a_ = ids_tensor((self.batch_size, 3) , config.vocab_size) a_ = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and a_ = torch.cat([input_ids, next_tokens] , dim=-1) a_ = torch.cat([input_mask, next_mask] , dim=-1) a_ = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , )["hidden_states"][0] a_ = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , )["hidden_states"][0] # select random slice a_ = ids_tensor((1,) , output_from_past.shape[-1]).item() a_ = output_from_no_past[:, -3:, random_slice_idx].detach() a_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , *__UpperCAmelCase , ) ->Tuple: a_ = BertGenerationDecoder(__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , labels=__UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase__ ( self) ->str: a_ , a_ , a_ , a_ = self.prepare_config_and_inputs() a_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): a_ : List[str] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () a_ : Optional[int] = (BertGenerationDecoder,) if is_torch_available() else () a_ : List[Any] = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def UpperCAmelCase__ ( self) ->List[Any]: a_ = BertGenerationEncoderTester(self) a_ = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37) def UpperCAmelCase__ ( self) ->Optional[Any]: self.config_tester.run_common_tests() def UpperCAmelCase__ ( self) ->Tuple: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->Tuple: a_ , a_ , a_ , a_ = self.model_tester.prepare_config_and_inputs() a_ = "bert" self.model_tester.create_and_check_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) def UpperCAmelCase__ ( self) ->int: a_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->List[str]: a_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->Optional[int]: # This regression test was failing with PyTorch < 1.3 ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() a_ = None self.model_tester.create_and_check_model_as_decoder( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) def UpperCAmelCase__ ( self) ->List[Any]: a_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*__UpperCAmelCase) @slow def UpperCAmelCase__ ( self) ->str: a_ = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder") self.assertIsNotNone(__UpperCAmelCase) @require_torch class snake_case ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self) ->int: a_ = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder") a_ = torch.tensor([[1_01, 75_92, 10_10, 20_26, 38_99, 20_03, 1_01_40, 1_02]]) with torch.no_grad(): a_ = model(__UpperCAmelCase)[0] a_ = torch.Size([1, 8, 10_24]) self.assertEqual(output.shape , __UpperCAmelCase) a_ = torch.tensor( [[[0.1_775, 0.0_083, -0.0_321], [1.6_002, 0.1_287, 0.3_912], [2.1_473, 0.5_791, 0.6_066]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4)) @require_torch class snake_case ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self) ->List[str]: a_ = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder") a_ = torch.tensor([[1_01, 75_92, 10_10, 20_26, 38_99, 20_03, 1_01_40, 1_02]]) with torch.no_grad(): a_ = model(__UpperCAmelCase)[0] a_ = torch.Size([1, 8, 5_03_58]) self.assertEqual(output.shape , __UpperCAmelCase) a_ = torch.tensor( [[[-0.5_788, -2.5_994, -3.7_054], [0.0_438, 4.7_997, 1.8_795], [1.5_862, 6.6_409, 4.4_638]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4))
303
1
import numpy # List of input, output pairs __lowerCamelCase : List[Any] = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) __lowerCamelCase : Dict = (((5_15, 22, 13), 5_55), ((61, 35, 49), 1_50)) __lowerCamelCase : List[str] = [2, 4, 1, 5] __lowerCamelCase : int = len(train_data) __lowerCamelCase : str = 0.009 def _snake_case ( lowerCAmelCase : Optional[Any] , lowerCAmelCase : Dict="train" ): """simple docstring""" return calculate_hypothesis_value(lowerCAmelCase , lowerCAmelCase ) - output( lowerCAmelCase , lowerCAmelCase ) def _snake_case ( lowerCAmelCase : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = 0 for i in range(len(lowerCAmelCase ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def _snake_case ( lowerCAmelCase : Union[str, Any] , lowerCAmelCase : List[str] ): """simple docstring""" if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def _snake_case ( lowerCAmelCase : Tuple , lowerCAmelCase : Tuple ): """simple docstring""" if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def _snake_case ( lowerCAmelCase : List[str] , lowerCAmelCase : str=m ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = 0 for i in range(lowerCAmelCase ): if index == -1: summation_value += _error(lowerCAmelCase ) else: summation_value += _error(lowerCAmelCase ) * train_data[i][0][index] return summation_value def _snake_case ( lowerCAmelCase : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = summation_of_cost_derivative(lowerCAmelCase , lowerCAmelCase ) / m return cost_derivative_value def _snake_case ( ): """simple docstring""" global parameter_vector # Tune these values to set a tolerance value for predicted output SCREAMING_SNAKE_CASE_ : List[str] = 0.000002 SCREAMING_SNAKE_CASE_ : Any = 0 SCREAMING_SNAKE_CASE_ : str = 0 while True: j += 1 SCREAMING_SNAKE_CASE_ : Optional[Any] = [0, 0, 0, 0] for i in range(0 , len(lowerCAmelCase ) ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = get_cost_derivative(i - 1 ) SCREAMING_SNAKE_CASE_ : Optional[int] = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( lowerCAmelCase , lowerCAmelCase , atol=lowerCAmelCase , rtol=lowerCAmelCase , ): break SCREAMING_SNAKE_CASE_ : Union[str, Any] = temp_parameter_vector print(("Number of iterations:", j) ) def _snake_case ( ): """simple docstring""" for i in range(len(lowerCAmelCase ) ): print(("Actual output value:", output(lowerCAmelCase , "test" )) ) print(("Hypothesis output:", calculate_hypothesis_value(lowerCAmelCase , "test" )) ) if __name__ == "__main__": run_gradient_descent() print('''\nTesting gradient descent for a linear hypothesis function.\n''') test_gradient_descent()
18
'''simple docstring''' import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging __UpperCAmelCase =logging.get_logger(__name__) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> None: __lowerCamelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(UpperCamelCase__ ) == len(UpperCamelCase__ ), f"""{len(UpperCamelCase__ )} != {len(UpperCamelCase__ )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) __UpperCAmelCase ={ # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 1_2: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 1_1], 4: [0, 4, 8, 1_1], 6: [0, 2, 4, 7, 9, 1_1], 9: [0, 1, 2, 4, 5, 7, 9, 1_0, 1_1], 1_2: list(range(1_2)), }, 1_6: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 1_5], 3: [0, 8, 1_5], 4: [0, 5, 1_0, 1_5], 6: [0, 3, 6, 9, 1_2, 1_5], 8: [0, 2, 4, 6, 8, 1_0, 1_2, 1_5], 9: [0, 1, 3, 5, 7, 9, 1_1, 1_3, 1_5], 1_2: [0, 1, 2, 3, 4, 5, 6, 7, 9, 1_1, 1_3, 1_5], 1_6: list(range(1_6)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } __UpperCAmelCase ={ # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 1_2: {1: [1_1], 2: [5, 1_1], 3: [3, 7, 1_1], 6: [1, 3, 5, 8, 1_0, 1_1]}, 1_6: {1: [1_5], 4: [4, 9, 1_2, 1_5], 8: [1, 3, 5, 7, 9, 1_1, 1_3, 1_5]}, } def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Union[str, Any]: try: __lowerCamelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( f"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" f""" {n_student}""" ) return list(range(UpperCamelCase__ ) ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> List[int]: if n_student > n_teacher: raise ValueError(f"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(UpperCamelCase__ ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ = "student" , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__=False , UpperCamelCase__=None , UpperCamelCase__=None , **UpperCamelCase__ , ) -> Tuple[PreTrainedModel, List[int], List[int]]: __lowerCamelCase = '''encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.''' assert (e is not None) or (d is not None), _msg if isinstance(UpperCamelCase__ , UpperCamelCase__ ): AutoTokenizer.from_pretrained(UpperCamelCase__ ).save_pretrained(UpperCamelCase__ ) # purely for convenience __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(UpperCamelCase__ ).eval() else: assert isinstance(UpperCamelCase__ , UpperCamelCase__ ), f"""teacher must be a model or string got type {type(UpperCamelCase__ )}""" __lowerCamelCase = teacher.config.to_diff_dict() try: __lowerCamelCase , __lowerCamelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __lowerCamelCase = teacher_e if d is None: __lowerCamelCase = teacher_d init_kwargs.update({'''encoder_layers''': e, '''decoder_layers''': d} ) except AttributeError: # T5 if hasattr(teacher.config , '''num_encoder_layers''' ): __lowerCamelCase , __lowerCamelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __lowerCamelCase , __lowerCamelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __lowerCamelCase = teacher_e if d is None: __lowerCamelCase = teacher_d if hasattr(teacher.config , '''num_encoder_layers''' ): init_kwargs.update({'''num_encoder_layers''': e, '''num_decoder_layers''': d} ) else: init_kwargs.update({'''num_layers''': e, '''num_decoder_layers''': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(UpperCamelCase__ ) # Copy weights __lowerCamelCase = teacher.config_class(**UpperCamelCase__ ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_config(UpperCamelCase__ ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __lowerCamelCase = student.load_state_dict(teacher.state_dict() , strict=UpperCamelCase__ ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __lowerCamelCase , __lowerCamelCase = list(range(UpperCamelCase__ ) ), list(range(UpperCamelCase__ ) ) logger.info( f"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" f""" {save_path}""" ) student.save_pretrained(UpperCamelCase__ ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __lowerCamelCase = pick_layers_to_copy(UpperCamelCase__ , UpperCamelCase__ ) if d_layers_to_copy is None: __lowerCamelCase = pick_layers_to_copy(UpperCamelCase__ , UpperCamelCase__ ) try: if hasattr( UpperCamelCase__ , '''prophetnet''' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , UpperCamelCase__ ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , UpperCamelCase__ ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , UpperCamelCase__ ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , UpperCamelCase__ ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , UpperCamelCase__ ) copy_layers(teacher.decoder.block , student.decoder.block , UpperCamelCase__ ) logger.info( f"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) __lowerCamelCase = { '''teacher_type''': teacher.config.model_type, '''copied_encoder_layers''': e_layers_to_copy, '''copied_decoder_layers''': d_layers_to_copy, } student.save_pretrained(UpperCamelCase__ ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
67
0
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' @staticmethod def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase : ArgumentParser ): """simple docstring""" UpperCAmelCase__ = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=_UpperCAmelCase , default=_UpperCAmelCase , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=_UpperCAmelCase , help="""Name of the model to download""" ) download_parser.set_defaults(func=_UpperCAmelCase ) def __init__( self : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : str , _UpperCAmelCase : bool , _UpperCAmelCase : bool ): """simple docstring""" UpperCAmelCase__ = model UpperCAmelCase__ = cache UpperCAmelCase__ = force UpperCAmelCase__ = trust_remote_code def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
61
'''simple docstring''' UpperCAmelCase_ = [sum(int(c, 1_0) ** 2 for c in i.__str__()) for i in range(1_0_0_0_0_0)] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' UpperCAmelCase__ = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 100000] number //= 100000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution UpperCAmelCase_ = [None] * 1_0_0_0_0_0_0_0 UpperCAmelCase_ = True UpperCAmelCase_ = False def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase__ = chain(next_number(SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ = number_chain while number < 10000000: UpperCAmelCase__ = number_chain number *= 10 return number_chain def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int = 10000000 ): '''simple docstring''' for i in range(1 , SCREAMING_SNAKE_CASE__ ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() print(f"{solution() = }")
61
1
'''simple docstring''' import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated lowerCamelCase : Dict = collections.namedtuple("_Datasets", ["train", "validation", "test"]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ lowerCamelCase : int = "https://storage.googleapis.com/cvdf-datasets/mnist/" def _lowerCAmelCase ( _UpperCamelCase : Dict ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =numpy.dtype(numpy.uintaa ).newbyteorder('>' ) return numpy.frombuffer(bytestream.read(4 ) , dtype=_UpperCamelCase )[0] @deprecated(_UpperCamelCase , 'Please use tf.data to implement this functionality.' ) def _lowerCAmelCase ( _UpperCamelCase : Union[str, Any] ) -> List[Any]: """simple docstring""" print('Extracting' , f.name ) with gzip.GzipFile(fileobj=_UpperCamelCase ) as bytestream: _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) if magic != 20_51: raise ValueError( 'Invalid magic number %d in MNIST image file: %s' % (magic, f.name) ) _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =bytestream.read(rows * cols * num_images ) _SCREAMING_SNAKE_CASE =numpy.frombuffer(_UpperCamelCase , dtype=numpy.uinta ) _SCREAMING_SNAKE_CASE =data.reshape(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , 1 ) return data @deprecated(_UpperCamelCase , 'Please use tf.one_hot on tensors.' ) def _lowerCAmelCase ( _UpperCamelCase : int , _UpperCamelCase : int ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =labels_dense.shape[0] _SCREAMING_SNAKE_CASE =numpy.arange(_UpperCamelCase ) * num_classes _SCREAMING_SNAKE_CASE =numpy.zeros((num_labels, num_classes) ) _SCREAMING_SNAKE_CASE =1 return labels_one_hot @deprecated(_UpperCamelCase , 'Please use tf.data to implement this functionality.' ) def _lowerCAmelCase ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Tuple=False , _UpperCamelCase : Tuple=10 ) -> str: """simple docstring""" print('Extracting' , f.name ) with gzip.GzipFile(fileobj=_UpperCamelCase ) as bytestream: _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) if magic != 20_49: raise ValueError( 'Invalid magic number %d in MNIST label file: %s' % (magic, f.name) ) _SCREAMING_SNAKE_CASE =_readaa(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =bytestream.read(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =numpy.frombuffer(_UpperCamelCase , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(_UpperCamelCase , _UpperCamelCase ) return labels class A__ : @deprecated( _a , 'Please use alternatives such as official/mnist/_DataSet.py' ' from tensorflow/models.' , ) def __init__( self : Any , _a : Union[str, Any] , _a : Optional[int] , _a : Dict=False , _a : Tuple=False , _a : List[Any]=dtypes.floataa , _a : Optional[int]=True , _a : List[Any]=None , ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =random_seed.get_seed(_a ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) _SCREAMING_SNAKE_CASE =dtypes.as_dtype(_a ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError('Invalid image dtype %r, expected uint8 or float32' % dtype ) if fake_data: _SCREAMING_SNAKE_CASE =1_0000 _SCREAMING_SNAKE_CASE =one_hot else: assert ( images.shape[0] == labels.shape[0] ), f"images.shape: {images.shape} labels.shape: {labels.shape}" _SCREAMING_SNAKE_CASE =images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 _SCREAMING_SNAKE_CASE =images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. _SCREAMING_SNAKE_CASE =images.astype(numpy.floataa ) _SCREAMING_SNAKE_CASE =numpy.multiply(_a , 1.0 / 2_55.0 ) _SCREAMING_SNAKE_CASE =images _SCREAMING_SNAKE_CASE =labels _SCREAMING_SNAKE_CASE =0 _SCREAMING_SNAKE_CASE =0 @property def A ( self : Union[str, Any] ) -> Dict: '''simple docstring''' return self._images @property def A ( self : str ) -> List[str]: '''simple docstring''' return self._labels @property def A ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' return self._num_examples @property def A ( self : Optional[int] ) -> List[str]: '''simple docstring''' return self._epochs_completed def A ( self : Dict , _a : Dict , _a : Dict=False , _a : Tuple=True ) -> Optional[Any]: '''simple docstring''' if fake_data: _SCREAMING_SNAKE_CASE =[1] * 784 _SCREAMING_SNAKE_CASE =[1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_a )], [fake_label for _ in range(_a )], ) _SCREAMING_SNAKE_CASE =self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: _SCREAMING_SNAKE_CASE =numpy.arange(self._num_examples ) numpy.random.shuffle(_a ) _SCREAMING_SNAKE_CASE =self.images[perma] _SCREAMING_SNAKE_CASE =self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch _SCREAMING_SNAKE_CASE =self._num_examples - start _SCREAMING_SNAKE_CASE =self._images[start : self._num_examples] _SCREAMING_SNAKE_CASE =self._labels[start : self._num_examples] # Shuffle the data if shuffle: _SCREAMING_SNAKE_CASE =numpy.arange(self._num_examples ) numpy.random.shuffle(_a ) _SCREAMING_SNAKE_CASE =self.images[perm] _SCREAMING_SNAKE_CASE =self.labels[perm] # Start next epoch _SCREAMING_SNAKE_CASE =0 _SCREAMING_SNAKE_CASE =batch_size - rest_num_examples _SCREAMING_SNAKE_CASE =self._index_in_epoch _SCREAMING_SNAKE_CASE =self._images[start:end] _SCREAMING_SNAKE_CASE =self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size _SCREAMING_SNAKE_CASE =self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(_UpperCamelCase , 'Please write your own downloading logic.' ) def _lowerCAmelCase ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] ) -> int: """simple docstring""" if not gfile.Exists(_UpperCamelCase ): gfile.MakeDirs(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =os.path.join(_UpperCamelCase , _UpperCamelCase ) if not gfile.Exists(_UpperCamelCase ): urllib.request.urlretrieve(_UpperCamelCase , _UpperCamelCase ) # noqa: S310 with gfile.GFile(_UpperCamelCase ) as f: _SCREAMING_SNAKE_CASE =f.size() print('Successfully downloaded' , _UpperCamelCase , _UpperCamelCase , 'bytes.' ) return filepath @deprecated( _UpperCamelCase , 'Please use alternatives such as:' ' tensorflow_datasets.load(\'mnist\')' ) def _lowerCAmelCase ( _UpperCamelCase : Optional[Any] , _UpperCamelCase : str=False , _UpperCamelCase : str=False , _UpperCamelCase : Any=dtypes.floataa , _UpperCamelCase : Any=True , _UpperCamelCase : Any=50_00 , _UpperCamelCase : Optional[int]=None , _UpperCamelCase : List[Any]=DEFAULT_SOURCE_URL , ) -> Optional[Any]: """simple docstring""" if fake_data: def fake(): return _DataSet( [] , [] , fake_data=_UpperCamelCase , one_hot=_UpperCamelCase , dtype=_UpperCamelCase , seed=_UpperCamelCase ) _SCREAMING_SNAKE_CASE =fake() _SCREAMING_SNAKE_CASE =fake() _SCREAMING_SNAKE_CASE =fake() return _Datasets(train=_UpperCamelCase , validation=_UpperCamelCase , test=_UpperCamelCase ) if not source_url: # empty string check _SCREAMING_SNAKE_CASE =DEFAULT_SOURCE_URL _SCREAMING_SNAKE_CASE ='train-images-idx3-ubyte.gz' _SCREAMING_SNAKE_CASE ='train-labels-idx1-ubyte.gz' _SCREAMING_SNAKE_CASE ='t10k-images-idx3-ubyte.gz' _SCREAMING_SNAKE_CASE ='t10k-labels-idx1-ubyte.gz' _SCREAMING_SNAKE_CASE =_maybe_download( _UpperCamelCase , _UpperCamelCase , source_url + train_images_file ) with gfile.Open(_UpperCamelCase , 'rb' ) as f: _SCREAMING_SNAKE_CASE =_extract_images(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_maybe_download( _UpperCamelCase , _UpperCamelCase , source_url + train_labels_file ) with gfile.Open(_UpperCamelCase , 'rb' ) as f: _SCREAMING_SNAKE_CASE =_extract_labels(_UpperCamelCase , one_hot=_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_maybe_download( _UpperCamelCase , _UpperCamelCase , source_url + test_images_file ) with gfile.Open(_UpperCamelCase , 'rb' ) as f: _SCREAMING_SNAKE_CASE =_extract_images(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_maybe_download( _UpperCamelCase , _UpperCamelCase , source_url + test_labels_file ) with gfile.Open(_UpperCamelCase , 'rb' ) as f: _SCREAMING_SNAKE_CASE =_extract_labels(_UpperCamelCase , one_hot=_UpperCamelCase ) if not 0 <= validation_size <= len(_UpperCamelCase ): _SCREAMING_SNAKE_CASE =( 'Validation size should be between 0 and ' f"{len(_UpperCamelCase )}. Received: {validation_size}." ) raise ValueError(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =train_images[:validation_size] _SCREAMING_SNAKE_CASE =train_labels[:validation_size] _SCREAMING_SNAKE_CASE =train_images[validation_size:] _SCREAMING_SNAKE_CASE =train_labels[validation_size:] _SCREAMING_SNAKE_CASE ={'dtype': dtype, 'reshape': reshape, 'seed': seed} _SCREAMING_SNAKE_CASE =_DataSet(_UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_DataSet(_UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_DataSet(_UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ) return _Datasets(train=_UpperCamelCase , validation=_UpperCamelCase , test=_UpperCamelCase )
47
import unittest from transformers import DonutProcessor lowerCamelCase = '''naver-clova-ix/donut-base''' class _a ( unittest.TestCase): def UpperCAmelCase__( self : str )-> int: lowerCAmelCase__ : Any = DonutProcessor.from_pretrained(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[int] )-> List[Any]: lowerCAmelCase__ : Dict = { '''name''': '''John Doe''', '''age''': '''99''', '''city''': '''Atlanta''', '''state''': '''GA''', '''zip''': '''30301''', '''phone''': '''123-4567''', '''nicknames''': [{'''nickname''': '''Johnny'''}, {'''nickname''': '''JD'''}], } lowerCAmelCase__ : Any = ( '''<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>''' '''<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>''' '''<s_nicknames><s_nickname>Johnny</s_nickname>''' '''<sep/><s_nickname>JD</s_nickname></s_nicknames>''' ) lowerCAmelCase__ : str = self.processor.tokenajson(_SCREAMING_SNAKE_CASE ) self.assertDictEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )
131
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[int] = '''ZinengTang/tvlt-base''' a :List[str] = tempfile.mkdtemp() def SCREAMING_SNAKE_CASE__ ( self , **_lowerCamelCase ): return TvltImageProcessor.from_pretrained(self.checkpoint , **_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , **_lowerCamelCase ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[Any] = self.get_image_processor() a :List[Any] = self.get_feature_extractor() a :List[Any] = TvltProcessor(image_processor=_lowerCamelCase , feature_extractor=_lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) a :List[Any] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , _lowerCamelCase ) self.assertIsInstance(processor.image_processor , _lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Union[str, Any] = self.get_image_processor() a :Union[str, Any] = self.get_feature_extractor() a :Optional[Any] = TvltProcessor(image_processor=_lowerCamelCase , feature_extractor=_lowerCamelCase ) a :List[Any] = np.ones([1_2000] ) a :str = feature_extractor(_lowerCamelCase , return_tensors='''np''' ) a :str = processor(audio=_lowerCamelCase , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self ): a :int = self.get_image_processor() a :Dict = self.get_feature_extractor() a :List[Any] = TvltProcessor(image_processor=_lowerCamelCase , feature_extractor=_lowerCamelCase ) a :Dict = np.ones([3, 224, 224] ) a :Tuple = image_processor(_lowerCamelCase , return_tensors='''np''' ) a :Optional[int] = processor(images=_lowerCamelCase , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self ): a :Tuple = self.get_image_processor() a :Union[str, Any] = self.get_feature_extractor() a :List[Any] = TvltProcessor(image_processor=_lowerCamelCase , feature_extractor=_lowerCamelCase ) a :Optional[int] = np.ones([1_2000] ) a :Optional[int] = np.ones([3, 224, 224] ) a :Union[str, Any] = processor(audio=_lowerCamelCase , images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(_lowerCamelCase ): processor() def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[int] = self.get_image_processor() a :List[Any] = self.get_feature_extractor() a :List[Any] = TvltProcessor(image_processor=_lowerCamelCase , feature_extractor=_lowerCamelCase ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
371
from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Union[str, Any] = logging.get_logger(__name__) snake_case : List[str] = { '''google/efficientnet-b7''': '''https://huggingface.co/google/efficientnet-b7/resolve/main/config.json''', } class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = 'efficientnet' def __init__( self , _lowerCamelCase = 3 , _lowerCamelCase = 600 , _lowerCamelCase = 2.0 , _lowerCamelCase = 3.1 , _lowerCamelCase = 8 , _lowerCamelCase = [3, 3, 5, 3, 5, 5, 3] , _lowerCamelCase = [32, 16, 24, 40, 80, 112, 192] , _lowerCamelCase = [16, 24, 40, 80, 112, 192, 320] , _lowerCamelCase = [] , _lowerCamelCase = [1, 2, 2, 2, 1, 2, 1] , _lowerCamelCase = [1, 2, 2, 3, 3, 4, 1] , _lowerCamelCase = [1, 6, 6, 6, 6, 6, 6] , _lowerCamelCase = 0.25 , _lowerCamelCase = "swish" , _lowerCamelCase = 2560 , _lowerCamelCase = "mean" , _lowerCamelCase = 0.02 , _lowerCamelCase = 0.001 , _lowerCamelCase = 0.99 , _lowerCamelCase = 0.5 , _lowerCamelCase = 0.2 , **_lowerCamelCase , ): super().__init__(**_lowerCamelCase ) a :Optional[int] = num_channels a :List[str] = image_size a :int = width_coefficient a :Optional[Any] = depth_coefficient a :Any = depth_divisor a :Any = kernel_sizes a :Tuple = in_channels a :Union[str, Any] = out_channels a :Any = depthwise_padding a :Any = strides a :Optional[Any] = num_block_repeats a :Tuple = expand_ratios a :Dict = squeeze_expansion_ratio a :int = hidden_act a :Dict = hidden_dim a :Tuple = pooling_type a :Any = initializer_range a :Tuple = batch_norm_eps a :Optional[int] = batch_norm_momentum a :List[Any] = dropout_rate a :Optional[int] = drop_connect_rate a :Tuple = sum(_lowerCamelCase ) * 4 class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return 1e-5
281
0
"""simple docstring""" from __future__ import annotations import requests def _snake_case ( snake_case__ : str ): A = F'https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty' return requests.get(snake_case__ ).json() def _snake_case ( snake_case__ : int = 10 ): A = 'https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty' A = requests.get(snake_case__ ).json()[:max_stories] return [get_hackernews_story(snake_case__ ) for story_id in story_ids] def _snake_case ( snake_case__ : int = 10 ): A = hackernews_top_stories(snake_case__ ) return "\n".join('* [{title}]({url})'.format(**snake_case__ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
74
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: List[str] = '''gpt_bigcode''' _lowerCamelCase: List[Any] = ['''past_key_values'''] _lowerCamelCase: int = { '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[int] ,A_ : Dict=5_0257 ,A_ : Union[str, Any]=1024 ,A_ : str=768 ,A_ : Any=12 ,A_ : Any=12 ,A_ : Optional[int]=None ,A_ : Any="gelu_pytorch_tanh" ,A_ : List[str]=0.1 ,A_ : Optional[int]=0.1 ,A_ : List[str]=0.1 ,A_ : Tuple=1e-5 ,A_ : Optional[int]=0.02 ,A_ : List[str]=True ,A_ : Optional[Any]=True ,A_ : List[Any]=5_0256 ,A_ : Union[str, Any]=5_0256 ,A_ : int=True ,A_ : Optional[Any]=True ,A_ : Dict=True ,**A_ : Union[str, Any] ,) -> Union[str, Any]: A = vocab_size A = n_positions A = n_embd A = n_layer A = n_head A = n_inner A = activation_function A = resid_pdrop A = embd_pdrop A = attn_pdrop A = layer_norm_epsilon A = initializer_range A = scale_attn_weights A = use_cache A = attention_softmax_in_fpaa A = scale_attention_softmax_in_fpaa A = multi_query A = bos_token_id A = eos_token_id super().__init__(bos_token_id=A_ ,eos_token_id=A_ ,**A_ )
74
1
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : int , _lowerCamelCase : int) -> list[str]: '''simple docstring''' if partitions <= 0: raise ValueError("partitions must be a positive number!") if partitions > number_of_bytes: raise ValueError("partitions can not > number_of_bytes!") __UpperCamelCase : Any = number_of_bytes // partitions __UpperCamelCase : Any = [] for i in range(_lowerCamelCase): __UpperCamelCase : int = i * bytes_per_partition + 1 __UpperCamelCase : List[str] = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F'{start_bytes}-{end_bytes}') return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
151
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[Any] = { 'configuration_pix2struct': [ 'PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Pix2StructConfig', 'Pix2StructTextConfig', 'Pix2StructVisionConfig', ], 'processing_pix2struct': ['Pix2StructProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : str = ['Pix2StructImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = [ 'PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Pix2StructPreTrainedModel', 'Pix2StructForConditionalGeneration', 'Pix2StructVisionModel', 'Pix2StructTextModel', ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
151
1