code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Any ) -> str: # picklable for multiprocessing """simple docstring""" return x.sum() def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Union[str, Any] ) -> Tuple: # picklable for multiprocessing """simple docstring""" return i + 1 @dataclass class _SCREAMING_SNAKE_CASE : snake_case__ : Dict = 4_2 snake_case__ : Dict = 4_2 class _SCREAMING_SNAKE_CASE ( _a ): def _A ( self : str ): UpperCamelCase :int = {} UpperCamelCase :Tuple = [] UpperCamelCase :Any = 1 UpperCamelCase :str = [1, 2] UpperCamelCase :List[str] = {"""a""": 1, """b""": 2} UpperCamelCase :List[str] = {"""a""": [1, 2], """b""": [3, 4]} UpperCamelCase :Dict = {"""a""": {"""1""": 1}, """b""": 2} UpperCamelCase :List[Any] = {"""a""": 1, """b""": 2, """c""": 3, """d""": 4} UpperCamelCase :List[str] = {} UpperCamelCase :List[str] = [] UpperCamelCase :str = 2 UpperCamelCase :Dict = [2, 3] UpperCamelCase :List[Any] = {"""a""": 2, """b""": 3} UpperCamelCase :Dict = {"""a""": [2, 3], """b""": [4, 5]} UpperCamelCase :List[str] = {"""a""": {"""1""": 2}, """b""": 3} UpperCamelCase :List[Any] = {"""a""": 2, """b""": 3, """c""": 4, """d""": 5} self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ ) , snake_case_ ) UpperCamelCase :List[str] = 2 self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual(map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) , snake_case_ ) UpperCamelCase :Tuple = {"""a""": np.eye(2 ), """b""": np.zeros(3 ), """c""": np.ones(2 )} UpperCamelCase :Union[str, Any] = {"""a""": 2, """b""": 0, """c""": 2} UpperCamelCase :Optional[Any] = { """a""": np.eye(2 ).astype(snake_case_ ), """b""": np.zeros(3 ).astype(snake_case_ ), """c""": np.ones(2 ).astype(snake_case_ ), } self.assertEqual(map_nested(snake_case_ , snake_case_ , map_numpy=snake_case_ ) , snake_case_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(snake_case_ , snake_case_ , map_numpy=snake_case_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(snake_case_ , snake_case_ , map_numpy=snake_case_ , num_proc=snake_case_ ) , snake_case_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(snake_case_ , snake_case_ , map_numpy=snake_case_ , num_proc=snake_case_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(snake_case_ ): # can't pickle a local lambda map_nested(lambda __lowerCamelCase : x + 1 , snake_case_ , num_proc=snake_case_ ) def _A ( self : int ): UpperCamelCase :Tuple = {"""a""": 1, """b""": 2} UpperCamelCase :Dict = {"""a""": 3, """b""": 4} UpperCamelCase :List[str] = {"""a""": 5, """b""": 6} UpperCamelCase :Tuple = sorted([("""a""", (1, 3, 5)), ("""b""", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(snake_case_ , snake_case_ , snake_case_ ) ) , snake_case_ ) def _A ( self : int ): class _SCREAMING_SNAKE_CASE : snake_case__ : str = """bar""" UpperCamelCase :Tuple = Foo() self.assertEqual(foo.my_attr , """bar""" ) with temporary_assignment(snake_case_ , """my_attr""" , """BAR""" ): self.assertEqual(foo.my_attr , """BAR""" ) self.assertEqual(foo.my_attr , """bar""" ) @pytest.mark.parametrize( """iterable_length, num_proc, expected_num_proc""" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[Any] , __magic_name__ : Dict , __magic_name__ : List[Any] ) -> Optional[int]: """simple docstring""" with patch("""datasets.utils.py_utils._single_map_nested""" ) as mock_single_map_nested, patch( """datasets.parallel.parallel.Pool""" ) as mock_multiprocessing_pool: UpperCamelCase :Union[str, Any] = {f"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase :int = map_nested(lambda __magic_name__ : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class _SCREAMING_SNAKE_CASE ( _a ): @require_tf def _A ( self : Optional[int] ): import tensorflow as tf from tensorflow.keras import layers UpperCamelCase :Tuple = layers.Dense(2 ) def gen_random_output(): UpperCamelCase :Union[str, Any] = tf.random.uniform((1, 3) ) return model(snake_case_ ).numpy() with temp_seed(42 , set_tensorflow=snake_case_ ): UpperCamelCase :List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=snake_case_ ): UpperCamelCase :List[str] = gen_random_output() UpperCamelCase :Any = gen_random_output() np.testing.assert_equal(snake_case_ , snake_case_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def _A ( self : Union[str, Any] ): import torch def gen_random_output(): UpperCamelCase :List[str] = torch.nn.Linear(3 , 2 ) UpperCamelCase :Tuple = torch.rand(1 , 3 ) return model(snake_case_ ).detach().numpy() with temp_seed(42 , set_pytorch=snake_case_ ): UpperCamelCase :List[str] = gen_random_output() with temp_seed(42 , set_pytorch=snake_case_ ): UpperCamelCase :List[str] = gen_random_output() UpperCamelCase :List[str] = gen_random_output() np.testing.assert_equal(snake_case_ , snake_case_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def _A ( self : Optional[Any] ): def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase :Union[str, Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase :List[str] = gen_random_output() UpperCamelCase :Optional[int] = gen_random_output() np.testing.assert_equal(snake_case_ , snake_case_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("""input_data""" , [{}] ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str ) -> Union[str, Any]: """simple docstring""" UpperCamelCase :Any = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( """data, expected_output""" , [ ({}, []), ([], []), ("""foo""", ["""foo"""]), (["""foo""", """bar"""], ["""foo""", """bar"""]), ([["""foo""", """bar"""]], ["""foo""", """bar"""]), ([[["""foo"""], ["""bar"""]]], ["""foo""", """bar"""]), ([[["""foo"""], """bar"""]], ["""foo""", """bar"""]), ({"""a""": 1, """b""": 2}, [1, 2]), ({"""a""": [1, 2], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[1, 2]], """b""": [[3, 4]]}, [1, 2, 3, 4]), ({"""a""": [[1, 2]], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [[[3], [4]]]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [[3, 4]]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [3, [4]]}, [1, 2, 3, 4]), ({"""a""": {"""1""": 1}, """b""": 2}, [1, 2]), ({"""a""": {"""1""": [1]}, """b""": 2}, [1, 2]), ({"""a""": {"""1""": [1]}, """b""": [2]}, [1, 2]), ] , ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str , __magic_name__ : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase :Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def SCREAMING_SNAKE_CASE_ ( ) -> Any: """simple docstring""" UpperCamelCase :Optional[int] = A(x=1 , y="""foobar""" ) UpperCamelCase :Any = {"""x""": 1, """y""": """foobar"""} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase :str = {"""a""": {"""b""": A(x=10 , y="""foo""" )}, """c""": [A(x=20 , y="""bar""" )]} UpperCamelCase :Optional[int] = {"""a""": {"""b""": {"""x""": 10, """y""": """foo"""}}, """c""": [{"""x""": 20, """y""": """bar"""}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="""foo""" )] ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : List[str] ) -> Optional[int]: """simple docstring""" return text.split() def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Tuple ) -> Optional[int]: """simple docstring""" yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def SCREAMING_SNAKE_CASE_ ( ) -> Optional[Any]: """simple docstring""" with Pool(2 ) as pool: UpperCamelCase :Tuple = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"""text""": """hello there"""}] * 10 ) ) assert out.count("""hello""" ) == 10 assert out.count("""there""" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase :Optional[Any] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"""text""": """hello there"""}] * 10 ) ) assert out.count("""hello""" ) == 10 assert out.count("""there""" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase :str = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"""content""": """a"""}, {"""content""": """b"""}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("""a""" ) == 2 assert out.count("""b""" ) == 2 assert len(_lowerCAmelCase ) == 4
38
'''simple docstring''' from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = CustomTokenizer pass
35
0
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __lowerCAmelCase ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self: Optional[int] ): # clean up the VRAM after each test super().tearDown() gc.collect() def SCREAMING_SNAKE_CASE ( self: str ): lowercase , lowercase :Dict = FlaxControlNetModel.from_pretrained( "lllyasviel/sd-controlnet-canny" , from_pt=_lowercase , dtype=jnp.bfloataa ) lowercase , lowercase :List[str] = FlaxStableDiffusionControlNetPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , controlnet=_lowercase , from_pt=_lowercase , dtype=jnp.bfloataa ) lowercase :List[str] = controlnet_params lowercase :Dict = "bird" lowercase :Union[str, Any] = jax.device_count() lowercase :Optional[int] = pipe.prepare_text_inputs([prompts] * num_samples ) lowercase :Tuple = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png" ) lowercase :Dict = pipe.prepare_image_inputs([canny_image] * num_samples ) lowercase :Any = jax.random.PRNGKey(0 ) lowercase :int = jax.random.split(_lowercase , jax.device_count() ) lowercase :Optional[Any] = replicate(_lowercase ) lowercase :List[Any] = shard(_lowercase ) lowercase :int = shard(_lowercase ) lowercase :str = pipe( prompt_ids=_lowercase , image=_lowercase , params=_lowercase , prng_seed=_lowercase , num_inference_steps=50 , jit=_lowercase , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowercase :Optional[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowercase :Union[str, Any] = images[0, 2_53:2_56, 2_53:2_56, -1] lowercase :Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowercase :Tuple = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: Any ): lowercase , lowercase :Any = FlaxControlNetModel.from_pretrained( "lllyasviel/sd-controlnet-openpose" , from_pt=_lowercase , dtype=jnp.bfloataa ) lowercase , lowercase :Optional[Any] = FlaxStableDiffusionControlNetPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , controlnet=_lowercase , from_pt=_lowercase , dtype=jnp.bfloataa ) lowercase :Union[str, Any] = controlnet_params lowercase :Optional[int] = "Chef in the kitchen" lowercase :str = jax.device_count() lowercase :Dict = pipe.prepare_text_inputs([prompts] * num_samples ) lowercase :str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png" ) lowercase :List[str] = pipe.prepare_image_inputs([pose_image] * num_samples ) lowercase :Optional[int] = jax.random.PRNGKey(0 ) lowercase :Any = jax.random.split(_lowercase , jax.device_count() ) lowercase :Any = replicate(_lowercase ) lowercase :List[str] = shard(_lowercase ) lowercase :Union[str, Any] = shard(_lowercase ) lowercase :str = pipe( prompt_ids=_lowercase , image=_lowercase , params=_lowercase , prng_seed=_lowercase , num_inference_steps=50 , jit=_lowercase , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowercase :Optional[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowercase :Tuple = images[0, 2_53:2_56, 2_53:2_56, -1] lowercase :List[str] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowercase :Optional[int] = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
370
from typing import TYPE_CHECKING from ...utils import _LazyModule _UpperCAmelCase : Optional[int] = {"processing_wav2vec2_with_lm": ["Wav2Vec2ProcessorWithLM"]} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys _UpperCAmelCase : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
158
0
from __future__ import annotations __A = list[tuple[int, int]] __A = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __A = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right class __lowerCAmelCase : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) -> int: '''simple docstring''' __lowerCamelCase = pos_x __lowerCamelCase = pos_y __lowerCamelCase = (pos_y, pos_x) __lowerCamelCase = goal_x __lowerCamelCase = goal_y __lowerCamelCase = g_cost __lowerCamelCase = parent __lowerCamelCase = self.calculate_heuristic() def lowercase_ ( self ) -> float: '''simple docstring''' __lowerCamelCase = abs(self.pos_x - self.goal_x ) __lowerCamelCase = abs(self.pos_y - self.goal_y ) return dx + dy def __lt__( self , lowerCamelCase__ ) -> bool: '''simple docstring''' return self.f_cost < other.f_cost class __lowerCAmelCase : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , __SCREAMING_SNAKE_CASE ) __lowerCamelCase = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , __SCREAMING_SNAKE_CASE ) __lowerCamelCase = [self.start] __lowerCamelCase = [] __lowerCamelCase = False def lowercase_ ( self ) -> Path | None: '''simple docstring''' while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() __lowerCamelCase = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: __lowerCamelCase = True return self.retrace_path(__SCREAMING_SNAKE_CASE ) self.closed_nodes.append(__SCREAMING_SNAKE_CASE ) __lowerCamelCase = self.get_successors(__SCREAMING_SNAKE_CASE ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(__SCREAMING_SNAKE_CASE ) else: # retrieve the best current path __lowerCamelCase = self.open_nodes.pop(self.open_nodes.index(__SCREAMING_SNAKE_CASE ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(__SCREAMING_SNAKE_CASE ) else: self.open_nodes.append(__SCREAMING_SNAKE_CASE ) if not self.reached: return [self.start.pos] return None def lowercase_ ( self , lowerCamelCase__ ) -> list[Node]: '''simple docstring''' __lowerCamelCase = [] for action in delta: __lowerCamelCase = parent.pos_x + action[1] __lowerCamelCase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(__SCREAMING_SNAKE_CASE ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , __SCREAMING_SNAKE_CASE , ) ) return successors def lowercase_ ( self , lowerCamelCase__ ) -> Path: '''simple docstring''' __lowerCamelCase = node __lowerCamelCase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) __lowerCamelCase = current_node.parent path.reverse() return path if __name__ == "__main__": __A = (0, 0) __A = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) print("------") __A = GreedyBestFirst(init, goal) __A = greedy_bf.search() if path: for pos_x, pos_y in path: __A = 2 for elem in grid: print(elem)
90
"""simple docstring""" from __future__ import annotations from typing import Generic, TypeVar _SCREAMING_SNAKE_CASE : Optional[Any] = TypeVar('''T''') class a ( Generic[T] ): def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : T ) -> None: lowerCamelCase_ = data lowerCamelCase_ = self lowerCamelCase_ = 0 class a ( Generic[T] ): def __init__( self : Any ) -> None: # map from node name to the node object lowerCamelCase_ = {} def UpperCamelCase ( self : List[str] , __SCREAMING_SNAKE_CASE : T ) -> None: # create a new set with x as its member lowerCamelCase_ = DisjointSetTreeNode(__SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self : Optional[int] , __SCREAMING_SNAKE_CASE : T ) -> DisjointSetTreeNode[T]: # find the set x belongs to (with path-compression) lowerCamelCase_ = self.map[data] if elem_ref != elem_ref.parent: lowerCamelCase_ = self.find_set(elem_ref.parent.data ) return elem_ref.parent def UpperCamelCase ( self : str , __SCREAMING_SNAKE_CASE : DisjointSetTreeNode[T] , __SCREAMING_SNAKE_CASE : DisjointSetTreeNode[T] ) -> None: # helper function for union operation if nodea.rank > nodea.rank: lowerCamelCase_ = nodea else: lowerCamelCase_ = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def UpperCamelCase ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : T ) -> None: # merge 2 disjoint sets self.link(self.find_set(__SCREAMING_SNAKE_CASE ) , self.find_set(__SCREAMING_SNAKE_CASE ) ) class a ( Generic[T] ): def __init__( self : Optional[int] ) -> None: # connections: map from the node to the neighbouring nodes (with weights) lowerCamelCase_ = {} def UpperCamelCase ( self : Optional[int] , __SCREAMING_SNAKE_CASE : T ) -> None: # add a node ONLY if its not present in the graph if node not in self.connections: lowerCamelCase_ = {} def UpperCamelCase ( self : Tuple , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # add an edge with the given weight self.add_node(__SCREAMING_SNAKE_CASE ) self.add_node(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = weight lowerCamelCase_ = weight def UpperCamelCase ( self : List[Any] ) -> GraphUndirectedWeighted[T]: lowerCamelCase_ = [] lowerCamelCase_ = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda __SCREAMING_SNAKE_CASE : x[2] ) # creating the disjoint set lowerCamelCase_ = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(__SCREAMING_SNAKE_CASE ) # MST generation lowerCamelCase_ = 0 lowerCamelCase_ = 0 lowerCamelCase_ = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = edges[index] index += 1 lowerCamelCase_ = disjoint_set.find_set(__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = disjoint_set.find_set(__SCREAMING_SNAKE_CASE ) if parent_u != parent_v: num_edges += 1 graph.add_edge(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) disjoint_set.union(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return graph
183
0
def __lowerCamelCase ( __magic_name__ : int = 1_000_000 ): a__: int =limit + 1 a__: List[str] =[0] * limit for first_term in range(1 , __magic_name__ ): for n in range(__magic_name__ , __magic_name__ , __magic_name__ ): a__: Union[str, Any] =first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Optional[int] =sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"""{solution() = }""")
42
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class lowerCamelCase__ ( _a ): _lowerCAmelCase = 42 class lowerCamelCase__ ( _a , _a ): @register_to_config def __init__( self : Union[str, Any] , _a : int = 1_6 , _a : int = 8_8 , _a : Optional[int] = None , _a : Optional[int] = None , _a : int = 1 , _a : float = 0.0 , _a : int = 3_2 , _a : Optional[int] = None , _a : bool = False , _a : Optional[int] = None , _a : str = "geglu" , _a : bool = True , _a : bool = True , ): super().__init__() a__: List[Any] =num_attention_heads a__: Tuple =attention_head_dim a__: Dict =num_attention_heads * attention_head_dim a__: List[Any] =in_channels a__: Dict =torch.nn.GroupNorm(num_groups=_a , num_channels=_a , eps=1e-6 , affine=_a ) a__: str =nn.Linear(_a , _a ) # 3. Define transformers blocks a__: Optional[int] =nn.ModuleList( [ BasicTransformerBlock( _a , _a , _a , dropout=_a , cross_attention_dim=_a , activation_fn=_a , attention_bias=_a , double_self_attention=_a , norm_elementwise_affine=_a , ) for d in range(_a ) ] ) a__: Any =nn.Linear(_a , _a ) def _lowerCamelCase ( self : List[str] , _a : str , _a : Optional[Any]=None , _a : int=None , _a : int=None , _a : Optional[int]=1 , _a : Tuple=None , _a : bool = True , ): a__ , a__ , a__ , a__: int =hidden_states.shape a__: str =batch_frames // num_frames a__: Any =hidden_states a__: Optional[int] =hidden_states[None, :].reshape(_a , _a , _a , _a , _a ) a__: Union[str, Any] =hidden_states.permute(0 , 2 , 1 , 3 , 4 ) a__: Tuple =self.norm(_a ) a__: Union[str, Any] =hidden_states.permute(0 , 3 , 4 , 2 , 1 ).reshape(batch_size * height * width , _a , _a ) a__: Dict =self.proj_in(_a ) # 2. Blocks for block in self.transformer_blocks: a__: str =block( _a , encoder_hidden_states=_a , timestep=_a , cross_attention_kwargs=_a , class_labels=_a , ) # 3. Output a__: Any =self.proj_out(_a ) a__: Optional[int] =( hidden_states[None, None, :] .reshape(_a , _a , _a , _a , _a ) .permute(0 , 3 , 4 , 1 , 2 ) .contiguous() ) a__: Dict =hidden_states.reshape(_a , _a , _a , _a ) a__: List[str] =hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=_a )
42
1
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class A ( unittest.TestCase ): def __init__(self , lowerCAmelCase , lowerCAmelCase=1_3 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=9_9 , lowerCAmelCase=3_2 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=3_7 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=5_1_2 , lowerCAmelCase=1_6 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=4 , ): __lowercase= parent __lowercase= batch_size __lowercase= seq_length __lowercase= is_training __lowercase= use_attention_mask __lowercase= use_token_type_ids __lowercase= use_labels __lowercase= vocab_size __lowercase= hidden_size __lowercase= num_hidden_layers __lowercase= num_attention_heads __lowercase= intermediate_size __lowercase= hidden_act __lowercase= hidden_dropout_prob __lowercase= attention_probs_dropout_prob __lowercase= max_position_embeddings __lowercase= type_vocab_size __lowercase= type_sequence_label_size __lowercase= initializer_range __lowercase= num_choices def _A (self ): __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase= None if self.use_attention_mask: __lowercase= random_attention_mask([self.batch_size, self.seq_length] ) __lowercase= DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCAmelCase , ) return config, input_ids, attention_mask def _A (self ): __lowercase= self.prepare_config_and_inputs() __lowercase, __lowercase, __lowercase= config_and_inputs __lowercase= {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class A ( A_ , unittest.TestCase ): UpperCamelCase_ : Optional[int] =( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def _A (self ): __lowercase= FlaxDistilBertModelTester(self ) @slow def _A (self ): for model_class_name in self.all_model_classes: __lowercase= model_class_name.from_pretrained('distilbert-base-uncased' ) __lowercase= model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase ) @require_flax class A ( unittest.TestCase ): @slow def _A (self ): __lowercase= FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) __lowercase= np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowercase= np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) __lowercase= model(lowerCAmelCase , attention_mask=lowerCAmelCase )[0] __lowercase= (1, 1_1, 7_6_8) self.assertEqual(output.shape , lowerCAmelCase ) __lowercase= np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCAmelCase , atol=1E-4 ) )
295
def _lowerCamelCase( lowercase__ , lowercase__ = " " ) -> list: '''simple docstring''' __lowercase= [] __lowercase= 0 for index, char in enumerate(lowercase__ ): if char == separator: split_words.append(string[last_index:index] ) __lowercase= index + 1 elif index + 1 == len(lowercase__ ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
295
1
'''simple docstring''' import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) snake_case_ : str = logging.getLogger() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : List[str] = '''\n'''.join(__lowerCAmelCase ) Path(__lowerCAmelCase ).open('w' ).writelines(__lowerCAmelCase ) snake_case_ : Optional[int] = 'patrickvonplaten/t5-tiny-random' snake_case_ : Tuple = 'sshleifer/bart-tiny-random' snake_case_ : Dict = 'sshleifer/tiny-mbart' snake_case_ : List[Any] = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowercase__ ( lowerCamelCase__ ): def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[Any] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = Path(self.get_auto_remove_tmp_dir() ) / '''utest_input.source''' _UpperCamelCase : int = input_file_name.parent / '''utest_output.txt''' assert not output_file_name.exists() _UpperCamelCase : Tuple = [''' New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County.'''] _dump_articles(snake_case__ ,snake_case__ ) _UpperCamelCase : Union[str, Any] = str(Path(self.get_auto_remove_tmp_dir() ) / 'scores.json' ) _UpperCamelCase : Dict = '''translation_en_to_de''' if model == T5_TINY else '''summarization''' _UpperCamelCase : Optional[Any] = F'\n run_eval_search.py\n {model}\n {input_file_name}\n {output_file_name}\n --score_path {score_path}\n --task {task}\n --num_beams 2\n --length_penalty 2.0\n '.split() with patch.object(snake_case__ ,'argv' ,snake_case__ ): run_generate() assert Path(snake_case__ ).exists() # os.remove(Path(output_file_name)) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' self.run_eval_tester(snake_case__ ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def UpperCamelCase_ ( self : str ,lowerCamelCase__ : str ): '''simple docstring''' self.run_eval_tester(snake_case__ ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def UpperCamelCase_ ( self : str ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : Tuple = Path(self.get_auto_remove_tmp_dir() ) / '''utest_input.source''' _UpperCamelCase : str = input_file_name.parent / '''utest_output.txt''' assert not output_file_name.exists() _UpperCamelCase : Tuple = { '''en''': ['''Machine learning is great, isn\'t it?''', '''I like to eat bananas''', '''Tomorrow is another great day!'''], '''de''': [ '''Maschinelles Lernen ist großartig, oder?''', '''Ich esse gerne Bananen''', '''Morgen ist wieder ein toller Tag!''', ], } _UpperCamelCase : Any = Path(self.get_auto_remove_tmp_dir() ) _UpperCamelCase : int = str(tmp_dir / 'scores.json' ) _UpperCamelCase : Optional[Any] = str(tmp_dir / 'val.target' ) _dump_articles(snake_case__ ,text['en'] ) _dump_articles(snake_case__ ,text['de'] ) _UpperCamelCase : Union[str, Any] = '''translation_en_to_de''' if model == T5_TINY else '''summarization''' _UpperCamelCase : str = F'\n run_eval_search.py\n {model}\n {str(snake_case__ )}\n {str(snake_case__ )}\n --score_path {score_path}\n --reference_path {reference_path}\n --task {task}\n '.split() testargs.extend(['--search', 'num_beams=1:2 length_penalty=0.9:1.0'] ) with patch.object(snake_case__ ,'argv' ,snake_case__ ): with CaptureStdout() as cs: run_search() _UpperCamelCase : str = [''' num_beams | length_penalty''', model, '''Best score args'''] _UpperCamelCase : str = ['''Info'''] if "translation" in task: expected_strings.append('bleu' ) else: expected_strings.extend(snake_case__ ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(snake_case__ ).exists() os.remove(Path(snake_case__ ) )
362
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case_ : List[Any] = logging.get_logger(__name__) snake_case_ : str = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all BART models at https://huggingface.co/models?filter=bart snake_case_ : Union[str, Any] = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, } snake_case_ : Any = { 'facebook/bart-base': 1024, 'facebook/bart-large': 1024, 'facebook/bart-large-mnli': 1024, 'facebook/bart-large-cnn': 1024, 'facebook/bart-large-xsum': 1024, 'yjernite/bart_eli5': 1024, } @lru_cache() def A__ ( ): _UpperCamelCase : str = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) _UpperCamelCase : Any = bs[:] _UpperCamelCase : Union[str, Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCAmelCase_ ) cs.append(2**8 + n ) n += 1 _UpperCamelCase : Any = [chr(UpperCAmelCase_ ) for n in cs] return dict(zip(UpperCAmelCase_ , UpperCAmelCase_ ) ) def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Tuple = set() _UpperCamelCase : Dict = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _UpperCamelCase : Any = char return pairs class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[int] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any]="replace" ,lowerCamelCase__ : str="<s>" ,lowerCamelCase__ : str="</s>" ,lowerCamelCase__ : str="</s>" ,lowerCamelCase__ : Any="<s>" ,lowerCamelCase__ : Tuple="<unk>" ,lowerCamelCase__ : List[str]="<pad>" ,lowerCamelCase__ : Optional[Any]="<mask>" ,lowerCamelCase__ : Tuple=False ,**lowerCamelCase__ : List[str] ,): '''simple docstring''' _UpperCamelCase : Dict = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else bos_token _UpperCamelCase : Tuple = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else eos_token _UpperCamelCase : Optional[int] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else sep_token _UpperCamelCase : Tuple = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else cls_token _UpperCamelCase : Union[str, Any] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else unk_token _UpperCamelCase : List[str] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _UpperCamelCase : str = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,add_prefix_space=lowerCamelCase__ ,**lowerCamelCase__ ,) with open(lowerCamelCase__ ,encoding='utf-8' ) as vocab_handle: _UpperCamelCase : List[str] = json.load(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {v: k for k, v in self.encoder.items()} _UpperCamelCase : Optional[Any] = errors # how to handle errors in decoding _UpperCamelCase : Tuple = bytes_to_unicode() _UpperCamelCase : List[str] = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ ,encoding='utf-8' ) as merges_handle: _UpperCamelCase : Dict = merges_handle.read().split('\n' )[1:-1] _UpperCamelCase : str = [tuple(merge.split() ) for merge in bpe_merges] _UpperCamelCase : Dict = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) _UpperCamelCase : Tuple = {} _UpperCamelCase : List[Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _UpperCamelCase : Any = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return len(self.encoder ) def UpperCamelCase_ ( self : str ): '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Any ): '''simple docstring''' if token in self.cache: return self.cache[token] _UpperCamelCase : Dict = tuple(lowerCamelCase__ ) _UpperCamelCase : List[str] = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _UpperCamelCase : List[str] = min(lowerCamelCase__ ,key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ ,float('inf' ) ) ) if bigram not in self.bpe_ranks: break _UpperCamelCase , _UpperCamelCase : int = bigram _UpperCamelCase : Optional[int] = [] _UpperCamelCase : Dict = 0 while i < len(lowerCamelCase__ ): try: _UpperCamelCase : int = word.index(lowerCamelCase__ ,lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _UpperCamelCase : Dict = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _UpperCamelCase : int = tuple(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = new_word if len(lowerCamelCase__ ) == 1: break else: _UpperCamelCase : Any = get_pairs(lowerCamelCase__ ) _UpperCamelCase : int = ' '.join(lowerCamelCase__ ) _UpperCamelCase : List[Any] = word return word def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : Optional[Any] ): '''simple docstring''' _UpperCamelCase : int = [] for token in re.findall(self.pat ,lowerCamelCase__ ): _UpperCamelCase : int = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(' ' ) ) return bpe_tokens def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[Any] ): '''simple docstring''' return self.encoder.get(lowerCamelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : int ): '''simple docstring''' return self.decoder.get(lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : Any ): '''simple docstring''' _UpperCamelCase : Dict = ''.join(lowerCamelCase__ ) _UpperCamelCase : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' ,errors=self.errors ) return text def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : List[Any] = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _UpperCamelCase : Union[str, Any] = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase__ ,ensure_ascii=lowerCamelCase__ ) + '\n' ) _UpperCamelCase : Optional[Any] = 0 with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' ' Please check that the tokenizer is not corrupted!' ) _UpperCamelCase : int = token_index writer.write(' '.join(lowerCamelCase__ ) + '\n' ) index += 1 return vocab_file, merge_file def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] _UpperCamelCase : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ,lowerCamelCase__ : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ ,token_ids_a=lowerCamelCase__ ,already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : Tuple = [self.sep_token_id] _UpperCamelCase : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Tuple=False ,**lowerCamelCase__ : Optional[int] ): '''simple docstring''' _UpperCamelCase : Tuple = kwargs.pop('add_prefix_space' ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _UpperCamelCase : List[str] = ' ' + text return (text, kwargs)
236
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A : str = { "configuration_pegasus_x": ["PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP", "PegasusXConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ "PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST", "PegasusXForConditionalGeneration", "PegasusXModel", "PegasusXPreTrainedModel", ] if TYPE_CHECKING: from .configuration_pegasus_x import PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP, PegasusXConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pegasus_x import ( PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST, PegasusXForConditionalGeneration, PegasusXModel, PegasusXPreTrainedModel, ) else: import sys A : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
57
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A : Optional[int] = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Tuple = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys A : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
57
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _lowerCamelCase = { 'vocab_file': {'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'}, 'tokenizer_file': { 'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json' }, } _lowerCamelCase = {'mobilebert-uncased': 5_12} _lowerCamelCase = {} class a ( _A ): '''simple docstring''' lowerCAmelCase : str = VOCAB_FILES_NAMES lowerCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase : Optional[Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase : Optional[Any] = MobileBertTokenizer def __init__( self : int , __snake_case : Dict=None , __snake_case : Any=None , __snake_case : int=True , __snake_case : Tuple="[UNK]" , __snake_case : List[str]="[SEP]" , __snake_case : int="[PAD]" , __snake_case : Any="[CLS]" , __snake_case : List[str]="[MASK]" , __snake_case : Tuple=True , __snake_case : Union[str, Any]=None , **__snake_case : List[Any] , ): super().__init__( __snake_case , tokenizer_file=__snake_case , do_lower_case=__snake_case , unk_token=__snake_case , sep_token=__snake_case , pad_token=__snake_case , cls_token=__snake_case , mask_token=__snake_case , tokenize_chinese_chars=__snake_case , strip_accents=__snake_case , **__snake_case , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , __snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , __snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , __snake_case ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(__snake_case , normalizer_state.pop('''type''' ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**__snake_case ) UpperCAmelCase_ = do_lower_case def lowerCamelCase_ ( self : Optional[int] , __snake_case : Optional[int] , __snake_case : Optional[Any]=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase_ ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase_ ( self : Optional[Any] , __snake_case : str , __snake_case : Optional[str] = None ): UpperCAmelCase_ = self._tokenizer.model.save(__snake_case , name=__snake_case ) return tuple(__snake_case )
177
from __future__ import annotations _lowerCamelCase = list[list[int]] # assigning initial values to the grid _lowerCamelCase = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution _lowerCamelCase = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Matrix , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : int ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Matrix ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Matrix ) -> Matrix | None: if location := find_empty_location(__UpperCamelCase ): UpperCAmelCase_ , UpperCAmelCase_ = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): UpperCAmelCase_ = digit if sudoku(__UpperCamelCase ) is not None: return grid UpperCAmelCase_ = 0 return None def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Matrix ) -> None: for row in grid: for cell in row: print(__UpperCamelCase , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') _lowerCamelCase = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
177
1
'''simple docstring''' from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase( self ) -> Dict: lowercase__ : int = tf.convert_to_tensor( [ [ 8.2_2_2_0_9_9_1, # 3rd highest value; idx. 0 -0.5_6_2_0_0_4_4, 5.2_3_2_2_9_7_5_2, 4.0_3_8_6_3_9_3, -6.8_7_9_8_3_7_8, -0.5_4_7_8_5_8_0_2, -3.2_0_1_2_1_5_3, 2.9_2_7_7_7_1_7_6, 1.8_8_1_7_1_9_5_3, 7.3_5_3_4_1_2_7_6, # 5th highest value; idx. 9 8.4_3_2_0_7_8_3_3, # 2nd highest value; idx. 10 -9.8_5_7_1_1_8_3_6, -5.9_6_2_0_9_2_3_6, -1.1_3_0_3_9_1_6_1, -7.1_1_1_5_2_9_4, -0.8_3_6_9_6_3_3, -5.3_1_8_6_4_0_8, 7.0_6_4_2_7_4_0_7, 0.8_1_3_6_9_3_4_4, -0.8_2_0_2_3_8_1_7, -5.9_1_7_9_7_9_6, 0.5_8_8_1_3_4_4_3, -6.9_9_7_7_8_4_3_8, 4.7_1_5_5_1_1_8_9, -0.1_8_7_7_1_6_3_7, 7.4_4_0_2_0_7_5_9, # 4th highest value; idx. 25 9.3_8_4_5_0_9_8_7, # 1st highest value; idx. 26 2.1_2_6_6_2_9_4_1, -9.3_2_5_6_2_0_3_8, 2.3_5_6_5_2_5_2_2, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5_8_4_2_5_5_1_8, 4.5_3_1_3_9_2_3_8, -5.5_7_5_1_0_4_6_4, -6.2_8_0_3_0_6_9_9, -7.1_9_5_2_9_5_0_3, -4.0_2_1_2_2_5_5_1, 1.3_9_3_3_7_0_3_7, -6.0_6_7_0_7_0_5_7, 1.5_9_4_8_0_5_1_7, -9.6_4_3_1_1_9, 0.0_3_9_0_7_7_9_9, 0.6_7_2_3_1_7_6_2, -8.8_8_2_0_6_7_2_6, 6.2_7_1_1_5_9_2_2, # 4th highest value; idx. 13 2.2_8_5_2_0_7_2_3, 4.8_2_7_6_7_5_0_6, 4.3_0_4_2_1_3_6_8, 8.8_2_7_5_3_1_3, # 2nd highest value; idx. 17 5.4_4_0_2_9_9_5_8, # 5th highest value; idx. 18 -4.4_7_3_5_7_9_4, 7.3_8_5_7_9_5_3_6, # 3rd highest value; idx. 20 -2.9_1_0_5_1_6_6_3, 2.6_1_9_4_6_0_7_7, -2.5_6_7_4_7_6_2, -9.4_8_9_5_9_3_0_2, -4.0_2_9_2_2_6_4_5, -1.3_5_4_1_6_9_1_8, 9.6_7_7_0_2_3_2_3, # 1st highest value; idx. 27 -5.8_9_4_7_8_5_5_3, 1.8_5_3_7_0_4_6_7, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) lowercase__ : str = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above lowercase__ : Any = tf.convert_to_tensor( [8.2_2_2_0_9_9, 7.3_5_3_4_1_2_6, 8.4_3_2_0_7_8, 7.4_4_0_2_0_7_5, 9.3_8_4_5_1, 6.2_7_1_1_5_9, 8.8_2_7_5_3_1, 5.4_4_0_2_9_9_5, 7.3_8_5_7_9_5_6, 9.6_7_7_0_2_3] , dtype=tf.floataa , ) # expected non filtered values as noted above lowercase__ : List[Any] = tf_top_k_top_p_filtering(__lowerCAmelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) lowercase__ : List[Any] = output[output != -float('''inf''' )] lowercase__ : Union[str, Any] = tf.cast( tf.where(tf.not_equal(__lowerCAmelCase , tf.constant(-float('''inf''' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__lowerCAmelCase , __lowerCAmelCase , rtol=1E-12 ) tf.debugging.assert_equal(__lowerCAmelCase , __lowerCAmelCase ) @require_tf class UpperCAmelCase ( unittest.TestCase , a__ ): '''simple docstring''' if is_tf_available(): SCREAMING_SNAKE_CASE = { "AutoModelForCausalLM": TFAutoModelForCausalLM, "AutoModelForSpeechSeq2Seq": TFAutoModelForSpeechSeqaSeq, "AutoModelForSeq2SeqLM": TFAutoModelForSeqaSeqLM, "AutoModelForVision2Seq": TFAutoModelForVisionaSeq, "LogitsProcessorList": TFLogitsProcessorList, "MinLengthLogitsProcessor": TFMinLengthLogitsProcessor, "create_tensor_fn": tf.convert_to_tensor, "floats_tensor": floats_tensor, "return_tensors": "tf", } @slow def _lowerCAmelCase( self ) -> Any: # TF-only test: tf.saved_model export lowercase__ : Optional[int] = TFAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowercase__ : Union[str, Any] = 2 lowercase__ : List[str] = 2 class UpperCAmelCase ( tf.Module ): '''simple docstring''' def __init__( self , __lowerCAmelCase ) -> Union[str, Any]: super(__lowerCAmelCase , self ).__init__() lowercase__ : int = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='''input_ids''' ), tf.TensorSpec((None, input_length) , tf.intaa , name='''attention_mask''' ), ) , jit_compile=__lowerCAmelCase , ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> str: lowercase__ : Optional[Any] = self.model.generate( input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase , max_new_tokens=__lowerCAmelCase , return_dict_in_generate=__lowerCAmelCase , ) return {"sequences": outputs["sequences"]} lowercase__ : List[str] = [[2, 0], [102, 103]] lowercase__ : Optional[Any] = [[1, 0], [1, 1]] lowercase__ : Any = DummyModel(model=__lowerCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__lowerCAmelCase , __lowerCAmelCase , signatures={'''serving_default''': dummy_model.serving} ) lowercase__ : str = tf.saved_model.load(__lowerCAmelCase ).signatures['''serving_default'''] for batch_size in range(1 , len(__lowerCAmelCase ) + 1 ): lowercase__ : List[str] = { '''input_ids''': tf.constant(dummy_input_ids[:batch_size] ), '''attention_mask''': tf.constant(dummy_attention_masks[:batch_size] ), } lowercase__ : List[str] = serving_func(**__lowerCAmelCase )['''sequences'''] lowercase__ : Tuple = test_model.generate(**__lowerCAmelCase , max_new_tokens=__lowerCAmelCase ) tf.debugging.assert_equal(__lowerCAmelCase , __lowerCAmelCase ) @slow def _lowerCAmelCase( self ) -> Optional[Any]: # TF-only test: tf.saved_model export lowercase__ : Tuple = TFAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowercase__ : Union[str, Any] = 1 lowercase__ : List[str] = 2 class UpperCAmelCase ( tf.Module ): '''simple docstring''' def __init__( self , __lowerCAmelCase ) -> int: super(__lowerCAmelCase , self ).__init__() lowercase__ : List[str] = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='''input_ids''' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='''attention_mask''' ), ) , jit_compile=__lowerCAmelCase , ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> Tuple: lowercase__ : int = self.model.generate( input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase , max_new_tokens=__lowerCAmelCase , return_dict_in_generate=__lowerCAmelCase , ) return {"sequences": outputs["sequences"]} lowercase__ : Union[str, Any] = [[2], [102, 103]] lowercase__ : str = [[1], [1, 1]] lowercase__ : int = DummyModel(model=__lowerCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__lowerCAmelCase , __lowerCAmelCase , signatures={'''serving_default''': dummy_model.serving} ) lowercase__ : Dict = tf.saved_model.load(__lowerCAmelCase ).signatures['''serving_default'''] for input_row in range(len(__lowerCAmelCase ) ): lowercase__ : Tuple = { '''input_ids''': tf.constant([dummy_input_ids[input_row]] ), '''attention_mask''': tf.constant([dummy_attention_masks[input_row]] ), } lowercase__ : Dict = serving_func(**__lowerCAmelCase )['''sequences'''] lowercase__ : Dict = test_model.generate(**__lowerCAmelCase , max_new_tokens=__lowerCAmelCase ) tf.debugging.assert_equal(__lowerCAmelCase , __lowerCAmelCase ) @slow @require_tensorflow_text def _lowerCAmelCase( self ) -> List[Any]: # TF-only test: tf.saved_model export with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='''google/flan-t5-small''' , filename='''spiece.model''' , local_dir=__lowerCAmelCase ) class UpperCAmelCase ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self ) -> Any: super().__init__() lowercase__ : List[Any] = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__lowerCAmelCase , '''spiece.model''' ) , '''rb''' ).read() ) lowercase__ : Any = TFAutoModelForSeqaSeqLM.from_pretrained('''hf-internal-testing/tiny-random-t5''' ) def _lowerCAmelCase( self , __lowerCAmelCase , *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: lowercase__ : Optional[int] = self.tokenizer.tokenize(__lowerCAmelCase ) lowercase__ , lowercase__ : Union[str, Any] = text.pad_model_inputs( __lowerCAmelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) lowercase__ : Optional[Any] = self.model.generate(input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase ) return self.tokenizer.detokenize(__lowerCAmelCase ) lowercase__ : List[Any] = CompleteSentenceTransformer() lowercase__ : Tuple = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='''inputs''' ) lowercase__ : Tuple = complete_model(__lowerCAmelCase ) lowercase__ : Optional[int] = tf.keras.Model(__lowerCAmelCase , __lowerCAmelCase ) keras_model.save(__lowerCAmelCase ) def _lowerCAmelCase( self ) -> Optional[int]: # Has PT equivalent: this test relies on random sampling lowercase__ : int = { '''do_sample''': True, '''num_beams''': 1, '''top_p''': 0.7, '''top_k''': 10, '''temperature''': 0.7, } lowercase__ : List[str] = 14 lowercase__ : str = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowercase__ : Any = '''Hello, my dog is cute and''' lowercase__ : Union[str, Any] = tokenizer(__lowerCAmelCase , return_tensors='''tf''' ) lowercase__ : List[str] = TFAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowercase__ : str = 638 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(''':/CPU:0''' ): tf.random.set_seed(0 ) lowercase__ : int = model.generate(**__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) lowercase__ : Any = [638, 198] with tf.device(''':/CPU:0''' ): tf.random.set_seed(0 ) lowercase__ : List[Any] = model.generate(**__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def _lowerCAmelCase( self ) -> List[str]: # Has PT equivalent: ample use of framework-specific code lowercase__ : Optional[int] = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bart''' ) lowercase__ : Union[str, Any] = '''Hugging Face is a technology company based in New York and Paris.''' lowercase__ : List[Any] = bart_tokenizer(__lowerCAmelCase , return_tensors='''tf''' ).input_ids lowercase__ : Tuple = TFBartForConditionalGeneration.from_pretrained('''hf-internal-testing/tiny-random-bart''' ) lowercase__ : Union[str, Any] = bart_model.generate(__lowerCAmelCase ).numpy() class UpperCAmelCase ( a__ ): '''simple docstring''' def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=None , **__lowerCAmelCase ) -> Union[str, Any]: return super().call(__lowerCAmelCase , **__lowerCAmelCase ) lowercase__ : str = FakeBart.from_pretrained('''hf-internal-testing/tiny-random-bart''' ) lowercase__ : Tuple = bart_model.generate(__lowerCAmelCase , foo='''bar''' ).numpy() self.assertTrue(np.array_equal(__lowerCAmelCase , __lowerCAmelCase ) ) class UpperCAmelCase ( bart_model.model.encoder.__class__ ): '''simple docstring''' def _lowerCAmelCase( self , __lowerCAmelCase , **__lowerCAmelCase ) -> Optional[int]: return super().call(__lowerCAmelCase , **__lowerCAmelCase ) lowercase__ : List[str] = FakeEncoder(bart_model.config , bart_model.model.shared ) lowercase__ : Optional[Any] = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) lowercase__ : Optional[int] = bart_model.generate(__lowerCAmelCase ).numpy() with self.assertRaises(__lowerCAmelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__lowerCAmelCase , foo='''bar''' )
198
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): lowercase__ : List[str] = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def __UpperCamelCase ( UpperCAmelCase = 100 ): lowercase__ : Dict = 1 lowercase__ : Optional[int] = 2 for i in range(2 , max_n + 1 ): lowercase__ : List[str] = pre_numerator lowercase__ : Optional[Any] = 2 * i // 3 if i % 3 == 0 else 1 lowercase__ : Union[str, Any] = cur_numerator lowercase__ : int = e_cont * pre_numerator + temp return sum_digits(UpperCAmelCase ) if __name__ == "__main__": print(F'{solution() = }')
198
1
from __future__ import annotations def _UpperCAmelCase ( _UpperCamelCase : List[Any], _UpperCamelCase : int, _UpperCamelCase : int ) -> tuple[float, list[float]]: A_ = list(range(len(__UpperCAmelCase ) ) ) A_ = [v / w for v, w in zip(__UpperCAmelCase, __UpperCAmelCase )] index.sort(key=lambda _UpperCamelCase : ratio[i], reverse=__UpperCAmelCase ) A_ = 0 A_ = [0] * len(__UpperCAmelCase ) for i in index: if weight[i] <= capacity: A_ = 1 max_value += value[i] capacity -= weight[i] else: A_ = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
355
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
18
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class _lowerCamelCase( a_ ): lowercase_ : List[Any] = ["""pixel_values"""] def __init__( self, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = PILImageResampling.BICUBIC, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = True, lowerCamelCase = 1 / 2_55, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = True, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" super().__init__(**_a) _lowercase : int = size if size is not None else {"shortest_edge": 2_24} _lowercase : Optional[Any] = get_size_dict(_a, default_to_square=_a) _lowercase : Optional[int] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} _lowercase : int = get_size_dict(_a, default_to_square=_a, param_name='crop_size') _lowercase : Optional[int] = do_resize _lowercase : str = size _lowercase : Any = resample _lowercase : Tuple = do_center_crop _lowercase : Dict = crop_size _lowercase : Dict = do_rescale _lowercase : List[Any] = rescale_factor _lowercase : str = do_normalize _lowercase : Dict = image_mean if image_mean is not None else OPENAI_CLIP_MEAN _lowercase : Optional[Any] = image_std if image_std is not None else OPENAI_CLIP_STD _lowercase : List[Any] = do_convert_rgb def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = PILImageResampling.BICUBIC, lowerCamelCase = None, **lowerCamelCase, ) -> Optional[Any]: """simple docstring""" _lowercase : Dict = get_size_dict(_a, default_to_square=_a) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''') _lowercase : Union[str, Any] = get_resize_output_image_size(_a, size=size['shortest_edge'], default_to_square=_a) return resize(_a, size=_a, resample=_a, data_format=_a, **_a) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = None, **lowerCamelCase, ) -> Dict: """simple docstring""" _lowercase : Optional[Any] = get_size_dict(_a) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''') return center_crop(_a, size=(size['height'], size['width']), data_format=_a, **_a) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = None, **lowerCamelCase, ) -> Dict: """simple docstring""" return rescale(_a, scale=_a, data_format=_a, **_a) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase = None, **lowerCamelCase, ) -> Optional[Any]: """simple docstring""" return normalize(_a, mean=_a, std=_a, data_format=_a, **_a) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase = ChannelDimension.FIRST, **lowerCamelCase, ) -> Tuple: """simple docstring""" _lowercase : Optional[int] = do_resize if do_resize is not None else self.do_resize _lowercase : Optional[int] = size if size is not None else self.size _lowercase : int = get_size_dict(_a, param_name='size', default_to_square=_a) _lowercase : str = resample if resample is not None else self.resample _lowercase : str = do_center_crop if do_center_crop is not None else self.do_center_crop _lowercase : List[Any] = crop_size if crop_size is not None else self.crop_size _lowercase : Optional[int] = get_size_dict(_a, param_name='crop_size', default_to_square=_a) _lowercase : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale _lowercase : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor _lowercase : Any = do_normalize if do_normalize is not None else self.do_normalize _lowercase : str = image_mean if image_mean is not None else self.image_mean _lowercase : Optional[Any] = image_std if image_std is not None else self.image_std _lowercase : Tuple = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb _lowercase : Optional[Any] = make_list_of_images(_a) if not valid_images(_a): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.') if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.') if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.') if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.') # PIL RGBA images are converted to RGB if do_convert_rgb: _lowercase : Any = [convert_to_rgb(_a) for image in images] # All transformations expect numpy arrays. _lowercase : str = [to_numpy_array(_a) for image in images] if do_resize: _lowercase : Union[str, Any] = [self.resize(image=_a, size=_a, resample=_a) for image in images] if do_center_crop: _lowercase : List[Any] = [self.center_crop(image=_a, size=_a) for image in images] if do_rescale: _lowercase : Dict = [self.rescale(image=_a, scale=_a) for image in images] if do_normalize: _lowercase : int = [self.normalize(image=_a, mean=_a, std=_a) for image in images] _lowercase : str = [to_channel_dimension_format(_a, _a) for image in images] _lowercase : List[Any] = {"pixel_values": images} return BatchFeature(data=_a, tensor_type=_a)
21
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class a__ : def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=32 , _a=2 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.0_2 , _a=3 , _a=4 , _a=None , _a=1_000 , ): lowercase : Optional[Any] = parent lowercase : Dict = batch_size lowercase : str = seq_length lowercase : List[Any] = is_training lowercase : Dict = use_input_mask lowercase : str = use_token_type_ids lowercase : int = use_labels lowercase : Union[str, Any] = vocab_size lowercase : Dict = hidden_size lowercase : List[str] = num_hidden_layers lowercase : Optional[int] = num_attention_heads lowercase : Tuple = intermediate_size lowercase : List[str] = hidden_act lowercase : int = hidden_dropout_prob lowercase : Any = attention_probs_dropout_prob lowercase : Dict = max_position_embeddings lowercase : Optional[int] = type_vocab_size lowercase : Tuple = type_sequence_label_size lowercase : Optional[int] = initializer_range lowercase : Dict = num_labels lowercase : Optional[int] = num_choices lowercase : List[Any] = scope lowercase : Dict = range_bbox def __magic_name__ ( self ): lowercase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment lowercase : int = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowercase : Any = bbox[i, j, 3] lowercase : Optional[Any] = bbox[i, j, 1] lowercase : Optional[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: lowercase : Dict = bbox[i, j, 2] lowercase : List[str] = bbox[i, j, 0] lowercase : List[Any] = t lowercase : Any = tf.convert_to_tensor(_a ) lowercase : Dict = None if self.use_input_mask: lowercase : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase : Optional[int] = None if self.use_token_type_ids: lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase : Optional[int] = None lowercase : List[Any] = None lowercase : Tuple = None if self.use_labels: lowercase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) lowercase : int = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __magic_name__ ( self , _a , _a , _a , _a , _a , _a , _a , _a ): lowercase : str = TFLayoutLMModel(config=_a ) lowercase : Optional[Any] = model(_a , _a , attention_mask=_a , token_type_ids=_a ) lowercase : Dict = model(_a , _a , token_type_ids=_a ) lowercase : List[str] = model(_a , _a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __magic_name__ ( self , _a , _a , _a , _a , _a , _a , _a , _a ): lowercase : List[Any] = TFLayoutLMForMaskedLM(config=_a ) lowercase : Union[str, Any] = model(_a , _a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __magic_name__ ( self , _a , _a , _a , _a , _a , _a , _a , _a ): lowercase : Dict = self.num_labels lowercase : Any = TFLayoutLMForSequenceClassification(config=_a ) lowercase : List[Any] = model(_a , _a , attention_mask=_a , token_type_ids=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __magic_name__ ( self , _a , _a , _a , _a , _a , _a , _a , _a ): lowercase : int = self.num_labels lowercase : Dict = TFLayoutLMForTokenClassification(config=_a ) lowercase : Tuple = model(_a , _a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __magic_name__ ( self , _a , _a , _a , _a , _a , _a , _a , _a ): lowercase : int = TFLayoutLMForQuestionAnswering(config=_a ) lowercase : Any = model(_a , _a , attention_mask=_a , token_type_ids=_a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __magic_name__ ( self ): lowercase : Optional[int] = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) : List[Any] = config_and_inputs lowercase : int = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class a__ ( a_, a_, unittest.TestCase ): __lowerCAmelCase = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __lowerCAmelCase = ( { """feature-extraction""": TFLayoutLMModel, """fill-mask""": TFLayoutLMForMaskedLM, """text-classification""": TFLayoutLMForSequenceClassification, """token-classification""": TFLayoutLMForTokenClassification, """zero-shot""": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = True __lowerCAmelCase = 10 def __magic_name__ ( self ): lowercase : List[Any] = TFLayoutLMModelTester(self ) lowercase : List[Any] = ConfigTester(self , config_class=_a , hidden_size=37 ) def __magic_name__ ( self ): self.config_tester.run_common_tests() def __magic_name__ ( self ): lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __magic_name__ ( self ): lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_a ) def __magic_name__ ( self ): lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_a ) def __magic_name__ ( self ): lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_a ) def __magic_name__ ( self ): lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_a ) @slow def __magic_name__ ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : List[str] = TFLayoutLMModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @unittest.skip("Onnx compliancy broke with TF 2.10" ) def __magic_name__ ( self ): pass def __magic_name__ ( ) -> Optional[int]: # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off lowercase : str = tf.convert_to_tensor([[101,1019,1014,1016,1037,1_2849,4747,1004,1_4246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,1_1300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,1_9274,2772,6205,2_7814,1_6147,1_6147,4343,2047,1_0283,1_0969,1_4389,1012,2338,102]] ) # noqa: E231 lowercase : Union[str, Any] = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 lowercase : Tuple = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 lowercase : Optional[Any] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) lowercase : List[Any] = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class a__ ( unittest.TestCase ): @slow def __magic_name__ ( self ): lowercase : Dict = TFLayoutLMModel.from_pretrained("microsoft/layoutlm-base-uncased" ) lowercase , lowercase , lowercase , lowercase , lowercase : Union[str, Any] = prepare_layoutlm_batch_inputs() # forward pass lowercase : Optional[int] = model(input_ids=_a , bbox=_a , attention_mask=_a , token_type_ids=_a ) # test the sequence output on [0, :3, :3] lowercase : Any = tf.convert_to_tensor( [[0.1_7_8_5, -0.1_9_4_7, -0.0_4_2_5], [-0.3_2_5_4, -0.2_8_0_7, 0.2_5_5_3], [-0.5_3_9_1, -0.3_3_2_2, 0.3_3_6_4]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , _a , atol=1E-3 ) ) # test the pooled output on [1, :3] lowercase : Optional[Any] = tf.convert_to_tensor([-0.6_5_8_0, -0.0_2_1_4, 0.8_5_5_2] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , _a , atol=1E-3 ) ) @slow def __magic_name__ ( self ): # initialize model with randomly initialized sequence classification head lowercase : List[Any] = TFLayoutLMForSequenceClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=2 ) lowercase , lowercase , lowercase , lowercase , lowercase : Optional[Any] = prepare_layoutlm_batch_inputs() # forward pass lowercase : Optional[Any] = model( input_ids=_a , bbox=_a , attention_mask=_a , token_type_ids=_a , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar lowercase : Union[str, Any] = outputs.loss lowercase : Union[str, Any] = (2,) self.assertEqual(loss.shape , _a ) # test the shape of the logits lowercase : List[str] = outputs.logits lowercase : Optional[Any] = (2, 2) self.assertEqual(logits.shape , _a ) @slow def __magic_name__ ( self ): # initialize model with randomly initialized token classification head lowercase : Any = TFLayoutLMForTokenClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=13 ) lowercase , lowercase , lowercase , lowercase , lowercase : str = prepare_layoutlm_batch_inputs() # forward pass lowercase : List[Any] = model( input_ids=_a , bbox=_a , attention_mask=_a , token_type_ids=_a , labels=_a ) # test the shape of the logits lowercase : int = outputs.logits lowercase : Optional[Any] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , _a ) @slow def __magic_name__ ( self ): # initialize model with randomly initialized token classification head lowercase : Union[str, Any] = TFLayoutLMForQuestionAnswering.from_pretrained("microsoft/layoutlm-base-uncased" ) lowercase , lowercase , lowercase , lowercase , lowercase : Tuple = prepare_layoutlm_batch_inputs() # forward pass lowercase : Optional[int] = model(input_ids=_a , bbox=_a , attention_mask=_a , token_type_ids=_a ) # test the shape of the logits lowercase : Any = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , _a ) self.assertEqual(outputs.end_logits.shape , _a )
202
0
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = len(_UpperCAmelCase ), len(grid[0] ) if ( min(_UpperCAmelCase , _UpperCAmelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowerCamelCase__ : Dict = 0 count += depth_first_search(_UpperCAmelCase , row + 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , row - 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col + 1 , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col - 1 , _UpperCAmelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
45
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class lowerCAmelCase ( pl.LightningModule ): def __init__( self : List[str] , UpperCAmelCase : Optional[Any] ) -> int: super().__init__() lowerCamelCase__ : List[str] = model lowerCamelCase__ : Dict = 2 lowerCamelCase__ : Dict = nn.Linear(self.model.config.hidden_size , self.num_labels ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: # load longformer model from model identifier lowerCamelCase__ : List[str] = LongformerModel.from_pretrained(_UpperCAmelCase ) lowerCamelCase__ : Dict = LightningModel(_UpperCAmelCase ) lowerCamelCase__ : List[Any] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lightning_model.load_state_dict(ckpt['state_dict'] ) # init longformer question answering model lowerCamelCase__ : Dict = LongformerForQuestionAnswering.from_pretrained(_UpperCAmelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(_UpperCAmelCase ) print(F"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _UpperCAmelCase : Any = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
45
1
"""simple docstring""" from typing import Any class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase ) -> List[Any]: _a = data _a = None class __lowerCamelCase : '''simple docstring''' def __init__( self ) -> Optional[Any]: _a = None def _UpperCAmelCase ( self ) -> Any: _a = self.head while temp is not None: print(temp.data , end=''' ''' ) _a = temp.next print() def _UpperCAmelCase ( self , __UpperCAmelCase ) -> Tuple: _a = Node(__UpperCAmelCase ) _a = self.head _a = new_node def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: if node_data_a == node_data_a: return else: _a = self.head while node_a is not None and node_a.data != node_data_a: _a = node_a.next _a = self.head while node_a is not None and node_a.data != node_data_a: _a = node_a.next if node_a is None or node_a is None: return _a = node_a.data, node_a.data if __name__ == "__main__": __snake_case = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
320
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline __A = "path-to-your-trained-model" __A = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("cuda") __A = "A photo of sks dog in a bucket" __A = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("dog-bucket.png")
148
0
'''simple docstring''' import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class _UpperCamelCase ( A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = MvpTokenizer lowerCAmelCase__ = MvpTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = filter_roberta_detectors def __lowerCamelCase ( self : List[Any]): '''simple docstring''' super().setUp() __lowercase =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __lowercase =dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase)))) __lowercase =['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __lowercase ={'unk_token': '<unk>'} __lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) __lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(_lowerCAmelCase) + '\n') with open(self.merges_file , 'w' , encoding='utf-8') as fp: fp.write('\n'.join(_lowerCAmelCase)) def __lowerCamelCase ( self : str , **_lowerCAmelCase : int): '''simple docstring''' kwargs.update(self.special_tokens_map) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase) def __lowerCamelCase ( self : Optional[int] , **_lowerCAmelCase : Any): '''simple docstring''' kwargs.update(self.special_tokens_map) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowerCAmelCase) def __lowerCamelCase ( self : str , _lowerCAmelCase : str): '''simple docstring''' return "lower newer", "lower newer" @cached_property def __lowerCamelCase ( self : List[str]): '''simple docstring''' return MvpTokenizer.from_pretrained('RUCAIBox/mvp') @cached_property def __lowerCamelCase ( self : List[str]): '''simple docstring''' return MvpTokenizerFast.from_pretrained('RUCAIBox/mvp') @require_torch def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =['A long paragraph for summarization.', 'Another paragraph for summarization.'] __lowercase =[0, 2_5_0, 2_5_1, 1_7_8_1_8, 1_3, 3_9_1_8_6, 1_9_3_8, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase =tokenizer(_lowerCAmelCase , max_length=len(_lowerCAmelCase) , padding=_lowerCAmelCase , return_tensors='pt') self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase) self.assertEqual((2, 9) , batch.input_ids.shape) self.assertEqual((2, 9) , batch.attention_mask.shape) __lowercase =batch.input_ids.tolist()[0] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase) # Test that special tokens are reset @require_torch def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' __lowercase =['A long paragraph for summarization.', 'Another paragraph for summarization.'] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase =tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors='pt') # check if input_ids are returned and no labels self.assertIn('input_ids' , _lowerCAmelCase) self.assertIn('attention_mask' , _lowerCAmelCase) self.assertNotIn('labels' , _lowerCAmelCase) self.assertNotIn('decoder_attention_mask' , _lowerCAmelCase) @require_torch def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase =[ 'Summary of the text.', 'Another summary.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase =tokenizer(text_target=_lowerCAmelCase , max_length=3_2 , padding='max_length' , return_tensors='pt') self.assertEqual(3_2 , targets['input_ids'].shape[1]) @require_torch def __lowerCamelCase ( self : int): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase =tokenizer( ['I am a small frog' * 1_0_2_4, 'I am a small frog'] , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors='pt') self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase) self.assertEqual(batch.input_ids.shape , (2, 1_0_2_4)) @require_torch def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __lowercase =['A long paragraph for summarization.'] __lowercase =[ 'Summary of the text.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase =tokenizer(_lowerCAmelCase , text_target=_lowerCAmelCase , return_tensors='pt') __lowercase =inputs['input_ids'] __lowercase =inputs['labels'] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item()) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item()) def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' pass def __lowerCamelCase ( self : Any): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})"""): __lowercase =self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase) __lowercase =self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase) __lowercase ='A, <mask> AllenNLP sentence.' __lowercase =tokenizer_r.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase) __lowercase =tokenizer_p.encode_plus(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids']) , sum(tokens_p['token_type_ids'])) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask']) / len(tokens_r['attention_mask']) , sum(tokens_p['attention_mask']) / len(tokens_p['attention_mask']) , ) __lowercase =tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids']) __lowercase =tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids']) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2]) self.assertSequenceEqual(tokens_r['input_ids'] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2]) self.assertSequenceEqual( _lowerCAmelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>']) self.assertSequenceEqual( _lowerCAmelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'])
48
'''simple docstring''' import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList lowerCamelCase = ["""\nclass""", """\ndef""", """\n#""", """\n@""", """\nprint""", """\nif"""] class _UpperCamelCase ( A ): '''simple docstring''' def __init__( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : Optional[int]=1): '''simple docstring''' __lowercase =tokenizer __lowercase =dataset __lowercase =len(_lowerCAmelCase) if n_tasks is None else n_tasks __lowercase =n_copies def __iter__( self : Union[str, Any]): '''simple docstring''' __lowercase =[] for task in range(self.n_tasks): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]['prompt'].strip()) __lowercase =self.tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors='pt') for task in range(self.n_tasks): for _ in range(self.n_copies): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class _UpperCamelCase ( A ): '''simple docstring''' def __init__( self : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any]): '''simple docstring''' __lowercase =start_length __lowercase =eof_strings __lowercase =tokenizer def __call__( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , **_lowerCAmelCase : Any): '''simple docstring''' __lowercase =self.tokenizer.batch_decode(input_ids[:, self.start_length :]) __lowercase =[] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings)) return all(_lowerCAmelCase) def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =re.split('(%s)' % '|'.join(_lowerCAmelCase ) , _lowerCAmelCase ) # last string should be "" return "".join(string_list[:-2] ) def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=20 , **_lowerCAmelCase ): """simple docstring""" __lowercase =defaultdict(_lowerCAmelCase ) # dict of list of generated tokens for step, batch in tqdm(enumerate(_lowerCAmelCase ) ): with torch.no_grad(): __lowercase =batch['ids'].shape[-1] __lowercase =accelerator.unwrap_model(_lowerCAmelCase ).generate( input_ids=batch['ids'][:, : batch['input_len']] , num_return_sequences=_lowerCAmelCase , **_lowerCAmelCase ) # each task is generated batch_size times __lowercase =batch['task_id'].repeat(_lowerCAmelCase ) __lowercase =accelerator.pad_across_processes( _lowerCAmelCase , dim=1 , pad_index=tokenizer.pad_token_id ) __lowercase , __lowercase =accelerator.gather((generated_tokens, generated_tasks) ) __lowercase =generated_tokens.cpu().numpy() __lowercase =generated_tasks.cpu().numpy() for task, generated_tokens in zip(_lowerCAmelCase , _lowerCAmelCase ): gen_token_dict[task].append(_lowerCAmelCase ) __lowercase =[[] for _ in range(_lowerCAmelCase )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: __lowercase =tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) code_gens[task].append(remove_last_block(_lowerCAmelCase ) ) return code_gens def _A ( ): """simple docstring""" __lowercase =HfArgumentParser(_lowerCAmelCase ) __lowercase =parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric __lowercase =args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing __lowercase ='false' if args.num_workers is None: __lowercase =multiprocessing.cpu_count() # Use dataset load to feed to accelerate __lowercase =Accelerator() set_seed(args.seed , device_specific=_lowerCAmelCase ) # Load model and tokenizer __lowercase =AutoTokenizer.from_pretrained(args.model_ckpt ) __lowercase =tokenizer.eos_token __lowercase =AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings __lowercase ={ 'do_sample': args.do_sample, 'temperature': args.temperature, 'max_new_tokens': args.max_new_tokens, 'top_p': args.top_p, 'top_k': args.top_k, 'stopping_criteria': StoppingCriteriaList([EndOfFunctionCriteria(0 , _lowerCAmelCase , _lowerCAmelCase )] ), } # Load evaluation dataset and metric __lowercase =load_dataset('openai_humaneval' ) __lowercase =load_metric('code_eval' ) __lowercase =args.num_tasks if args.num_tasks is not None else len(human_eval['test'] ) __lowercase =args.n_samples // args.batch_size __lowercase =TokenizedDataset(_lowerCAmelCase , human_eval['test'] , n_copies=_lowerCAmelCase , n_tasks=_lowerCAmelCase ) # do not confuse args.batch_size, which is actually the num_return_sequences __lowercase =DataLoader(_lowerCAmelCase , batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: __lowercase =code_eval_metric.compute(references=[''] , predictions=[['']] ) except ValueError as exception: print( 'Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL="1"`' ' flag to enable code evaluation.' ) raise exception __lowercase , __lowercase =accelerator.prepare(_lowerCAmelCase , _lowerCAmelCase ) __lowercase =complete_code( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , n_tasks=_lowerCAmelCase , batch_size=args.batch_size , **_lowerCAmelCase , ) if accelerator.is_main_process: __lowercase =[] for task in tqdm(range(_lowerCAmelCase ) ): __lowercase =human_eval['test'][task]['test'] __lowercase =f"""check({human_eval['test'][task]['entry_point']})""" references.append('\n' + test_func + '\n' + entry_point ) # Evaluate completions with "code_eval" metric __lowercase , __lowercase =code_eval_metric.compute( references=_lowerCAmelCase , predictions=_lowerCAmelCase , num_workers=args.num_workers ) print(f"""Results: {pass_at_k}""" ) # Save results to json file with open(args.output_file , 'w' ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
48
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer __UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name __UpperCAmelCase = '\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu")\n\n >>> repo = "openai/shap-e-img2img"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png"\n >>> image = load_image(image_url).convert("RGB")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], "corgi_3d.gif")\n ```\n' @dataclass class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Union[PIL.Image.Image, np.ndarray] class lowerCamelCase (_snake_case ): '''simple docstring''' def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Any: super().__init__() self.register_modules( prior=_UpperCamelCase , image_encoder=_UpperCamelCase , image_processor=_UpperCamelCase , scheduler=_UpperCamelCase , renderer=_UpperCamelCase , ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> List[Any]: if latents is None: UpperCAmelCase_ : str = randn_tensor(_UpperCamelCase , generator=_UpperCamelCase , device=_UpperCamelCase , dtype=_UpperCamelCase ) else: if latents.shape != shape: raise ValueError(f"Unexpected latents shape, got {latents.shape}, expected {shape}" ) UpperCAmelCase_ : Tuple = latents.to(_UpperCamelCase ) UpperCAmelCase_ : Tuple = latents * scheduler.init_noise_sigma return latents def __UpperCAmelCase ( self , _UpperCamelCase=0 ) -> Union[str, Any]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) UpperCAmelCase_ : int = torch.device(f"cuda:{gpu_id}" ) UpperCAmelCase_ : int = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_UpperCamelCase , _UpperCamelCase ) @property def __UpperCAmelCase ( self ) -> int: if self.device != torch.device('meta' ) or not hasattr(self.image_encoder , '_hf_hook' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(_UpperCamelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> str: if isinstance(_UpperCamelCase , _UpperCamelCase ) and isinstance(image[0] , torch.Tensor ): UpperCAmelCase_ : int = torch.cat(_UpperCamelCase , axis=0 ) if image[0].ndim == 4 else torch.stack(_UpperCamelCase , axis=0 ) if not isinstance(_UpperCamelCase , torch.Tensor ): UpperCAmelCase_ : Optional[int] = self.image_processor(_UpperCamelCase , return_tensors='pt' ).pixel_values[0].unsqueeze(0 ) UpperCAmelCase_ : Tuple = image.to(dtype=self.image_encoder.dtype , device=_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = self.image_encoder(_UpperCamelCase )['last_hidden_state'] UpperCAmelCase_ : Union[str, Any] = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 UpperCAmelCase_ : List[str] = image_embeds.repeat_interleave(_UpperCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Dict = torch.zeros_like(_UpperCamelCase ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase_ : Optional[int] = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(_UpperCamelCase ) def __call__( self , _UpperCamelCase , _UpperCamelCase = 1 , _UpperCamelCase = 2_5 , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = 4.0 , _UpperCamelCase = 6_4 , _UpperCamelCase = "pil" , _UpperCamelCase = True , ) -> Union[str, Any]: if isinstance(_UpperCamelCase , PIL.Image.Image ): UpperCAmelCase_ : Tuple = 1 elif isinstance(_UpperCamelCase , torch.Tensor ): UpperCAmelCase_ : str = image.shape[0] elif isinstance(_UpperCamelCase , _UpperCamelCase ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): UpperCAmelCase_ : Optional[int] = len(_UpperCamelCase ) else: raise ValueError( f"`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(_UpperCamelCase )}" ) UpperCAmelCase_ : Tuple = self._execution_device UpperCAmelCase_ : str = batch_size * num_images_per_prompt UpperCAmelCase_ : str = guidance_scale > 1.0 UpperCAmelCase_ : str = self._encode_image(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # prior self.scheduler.set_timesteps(_UpperCamelCase , device=_UpperCamelCase ) UpperCAmelCase_ : int = self.scheduler.timesteps UpperCAmelCase_ : int = self.prior.config.num_embeddings UpperCAmelCase_ : Any = self.prior.config.embedding_dim UpperCAmelCase_ : List[str] = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim UpperCAmelCase_ : List[Any] = latents.reshape(latents.shape[0] , _UpperCamelCase , _UpperCamelCase ) for i, t in enumerate(self.progress_bar(_UpperCamelCase ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Optional[Any] = self.scheduler.scale_model_input(_UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : int = self.prior( _UpperCamelCase , timestep=_UpperCamelCase , proj_embedding=_UpperCamelCase , ).predicted_image_embedding # remove the variance UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: UpperCAmelCase_ , UpperCAmelCase_ : str = noise_pred.chunk(2 ) UpperCAmelCase_ : List[Any] = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) UpperCAmelCase_ : List[str] = self.scheduler.step( _UpperCamelCase , timestep=_UpperCamelCase , sample=_UpperCamelCase , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=_UpperCamelCase ) UpperCAmelCase_ : List[Any] = [] for i, latent in enumerate(_UpperCamelCase ): print() UpperCAmelCase_ : List[str] = self.renderer.decode( latent[None, :] , _UpperCamelCase , size=_UpperCamelCase , ray_batch_size=4_0_9_6 , n_coarse_samples=6_4 , n_fine_samples=1_2_8 , ) images.append(_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = torch.stack(_UpperCamelCase ) if output_type not in ["np", "pil"]: raise ValueError(f"Only the output types `pil` and `np` are supported not output_type={output_type}" ) UpperCAmelCase_ : Dict = images.cpu().numpy() if output_type == "pil": UpperCAmelCase_ : List[str] = [self.numpy_to_pil(_UpperCamelCase ) for image in images] # Offload last model to CPU if hasattr(self , 'final_offload_hook' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=_UpperCamelCase )
29
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices a__ = logging.get_logger(__name__) a__ = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class snake_case ( SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' snake_case_ : Optional[Any] = """swin""" snake_case_ : Optional[Any] = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : str , lowerCAmelCase : Optional[int]=224 , lowerCAmelCase : int=4 , lowerCAmelCase : Any=3 , lowerCAmelCase : int=96 , lowerCAmelCase : Optional[Any]=[2, 2, 6, 2] , lowerCAmelCase : Optional[Any]=[3, 6, 12, 24] , lowerCAmelCase : Tuple=7 , lowerCAmelCase : List[Any]=4.0 , lowerCAmelCase : Tuple=True , lowerCAmelCase : Optional[int]=0.0 , lowerCAmelCase : Union[str, Any]=0.0 , lowerCAmelCase : Optional[int]=0.1 , lowerCAmelCase : Tuple="gelu" , lowerCAmelCase : Any=False , lowerCAmelCase : Union[str, Any]=0.02 , lowerCAmelCase : int=1E-5 , lowerCAmelCase : Optional[Any]=32 , lowerCAmelCase : Optional[int]=None , lowerCAmelCase : Dict=None , **lowerCAmelCase : Tuple , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowerCAmelCase) _snake_case : int = image_size _snake_case : Any = patch_size _snake_case : Union[str, Any] = num_channels _snake_case : int = embed_dim _snake_case : Dict = depths _snake_case : Dict = len(lowerCAmelCase) _snake_case : Optional[Any] = num_heads _snake_case : Tuple = window_size _snake_case : int = mlp_ratio _snake_case : Any = qkv_bias _snake_case : Union[str, Any] = hidden_dropout_prob _snake_case : List[str] = attention_probs_dropout_prob _snake_case : Optional[Any] = drop_path_rate _snake_case : List[Any] = hidden_act _snake_case : str = use_absolute_embeddings _snake_case : Tuple = layer_norm_eps _snake_case : Any = initializer_range _snake_case : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _snake_case : Dict = int(embed_dim * 2 ** (len(lowerCAmelCase) - 1)) _snake_case : Optional[Any] = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 , len(lowerCAmelCase) + 1)] _snake_case , _snake_case : List[str] = get_aligned_output_features_output_indices( out_features=lowerCAmelCase , out_indices=lowerCAmelCase , stage_names=self.stage_names) class snake_case ( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' snake_case_ : int = version.parse("""1.11""" ) @property def UpperCamelCase_ ( self : Dict) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ]) @property def UpperCamelCase_ ( self : Dict) -> float: """simple docstring""" return 1E-4
317
0
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _SCREAMING_SNAKE_CASE : def __init__( self : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : List[Any]=13 , __lowerCamelCase : Union[str, Any]=32 , __lowerCamelCase : List[Any]=3 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : Optional[int]=[10, 20, 30, 40] , __lowerCamelCase : int=[2, 2, 3, 2] , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : Tuple=37 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : str=10 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Tuple=["stage2", "stage3", "stage4"] , __lowerCamelCase : Tuple=[2, 3, 4] , __lowerCamelCase : Tuple=None , ): UpperCamelCase :Union[str, Any] = parent UpperCamelCase :List[str] = batch_size UpperCamelCase :Any = image_size UpperCamelCase :Any = num_channels UpperCamelCase :Union[str, Any] = num_stages UpperCamelCase :List[Any] = hidden_sizes UpperCamelCase :List[Any] = depths UpperCamelCase :List[str] = is_training UpperCamelCase :int = use_labels UpperCamelCase :Union[str, Any] = intermediate_size UpperCamelCase :List[Any] = hidden_act UpperCamelCase :Any = num_labels UpperCamelCase :Any = initializer_range UpperCamelCase :int = out_features UpperCamelCase :int = out_indices UpperCamelCase :Optional[Any] = scope def _A ( self : Any ): UpperCamelCase :Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase :Tuple = None if self.use_labels: UpperCamelCase :Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase :Any = self.get_config() return config, pixel_values, labels def _A ( self : int ): return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _A ( self : Optional[int] , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict ): UpperCamelCase :List[Any] = ConvNextModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Any = model(__lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _A ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] ): UpperCamelCase :Dict = ConvNextForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :str = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _A ( self : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any] ): UpperCamelCase :Optional[Any] = ConvNextBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = model(__lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCamelCase :Any = None UpperCamelCase :List[Any] = ConvNextBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = model(__lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _A ( self : str ): UpperCamelCase :str = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase :Optional[int] = config_and_inputs UpperCamelCase :Dict = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case__ : Tuple = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) snake_case__ : str = ( {"""feature-extraction""": ConvNextModel, """image-classification""": ConvNextForImageClassification} if is_torch_available() else {} ) snake_case__ : List[str] = True snake_case__ : Any = False snake_case__ : int = False snake_case__ : Union[str, Any] = False snake_case__ : Dict = False def _A ( self : Optional[int] ): UpperCamelCase :int = ConvNextModelTester(self ) UpperCamelCase :List[str] = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def _A ( self : Optional[int] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _A ( self : Any ): return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def _A ( self : Dict ): pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def _A ( self : Optional[int] ): pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def _A ( self : Tuple ): pass def _A ( self : Union[str, Any] ): UpperCamelCase , UpperCamelCase :str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :str = model_class(__lowerCamelCase ) UpperCamelCase :Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase :str = [*signature.parameters.keys()] UpperCamelCase :Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _A ( self : str ): UpperCamelCase :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__lowerCamelCase ) def _A ( self : Dict ): def check_hidden_states_output(__lowerCamelCase : List[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[int] ): UpperCamelCase :Any = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): UpperCamelCase :List[str] = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) UpperCamelCase :List[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase :Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCamelCase , UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :Any = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase :List[Any] = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _A ( self : Any ): UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _A ( self : Dict ): for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase :List[Any] = ConvNextModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def SCREAMING_SNAKE_CASE_ ( ) -> str: """simple docstring""" UpperCamelCase :Union[str, Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _A ( self : List[Any] ): return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def _A ( self : Optional[int] ): UpperCamelCase :Dict = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(__lowerCamelCase ) UpperCamelCase :int = self.default_image_processor UpperCamelCase :Dict = prepare_img() UpperCamelCase :Any = image_processor(images=__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): UpperCamelCase :Union[str, Any] = model(**__lowerCamelCase ) # verify the logits UpperCamelCase :str = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCamelCase :List[Any] = torch.tensor([-0.0260, -0.4739, 0.1911] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase , _a ): snake_case__ : List[str] = (ConvNextBackbone,) if is_torch_available() else () snake_case__ : int = ConvNextConfig snake_case__ : Dict = False def _A ( self : Any ): UpperCamelCase :Any = ConvNextModelTester(self )
62
import string def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str ) -> None: """simple docstring""" for key in range(len(string.ascii_uppercase ) ): UpperCamelCase :List[str] = """""" for symbol in message: if symbol in string.ascii_uppercase: UpperCamelCase :Optional[Any] = string.ascii_uppercase.find(__magic_name__ ) UpperCamelCase :Any = num - key if num < 0: UpperCamelCase :Optional[Any] = num + len(string.ascii_uppercase ) UpperCamelCase :int = translated + string.ascii_uppercase[num] else: UpperCamelCase :Optional[Any] = translated + symbol print(f"""Decryption using Key #{key}: {translated}""" ) def SCREAMING_SNAKE_CASE_ ( ) -> None: """simple docstring""" UpperCamelCase :List[Any] = input("""Encrypted message: """ ) UpperCamelCase :Optional[Any] = message.upper() decrypt(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
62
1
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowercase__ ( metaclass=lowercase ): lowercase__ = ["""note_seq"""] def __init__( self : Tuple ,*lowerCamelCase__ : Union[str, Any] ,**lowerCamelCase__ : Tuple ): '''simple docstring''' requires_backends(self ,['note_seq'] ) @classmethod def UpperCamelCase_ ( cls : List[str] ,*lowerCamelCase__ : Tuple ,**lowerCamelCase__ : int ): '''simple docstring''' requires_backends(cls ,['note_seq'] ) @classmethod def UpperCamelCase_ ( cls : str ,*lowerCamelCase__ : Union[str, Any] ,**lowerCamelCase__ : Any ): '''simple docstring''' requires_backends(cls ,['note_seq'] )
83
'''simple docstring''' import math import sys def __a(SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' if number != int(SCREAMING_SNAKE_CASE_ ): raise ValueError("the value of input must be a natural number" ) if number < 0: raise ValueError("the value of input must not be a negative number" ) if number == 0: return 1 _lowerCAmelCase = [-1] * (number + 1) _lowerCAmelCase = 0 for i in range(1 , number + 1 ): _lowerCAmelCase = sys.maxsize _lowerCAmelCase = int(math.sqrt(SCREAMING_SNAKE_CASE_ ) ) for j in range(1 , root + 1 ): _lowerCAmelCase = 1 + answers[i - (j**2)] _lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) _lowerCAmelCase = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
158
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) _lowerCAmelCase : Optional[int] = { '''tiiuae/falcon-40b''': '''https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json''', '''tiiuae/falcon-7b''': '''https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json''', } class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = '''falcon''' __UpperCamelCase = ['''past_key_values'''] def __init__( self :List[Any] , snake_case :Optional[int]=65_024 , snake_case :Tuple=4_544 , snake_case :Dict=32 , snake_case :Union[str, Any]=71 , snake_case :List[Any]=1e-5 , snake_case :Union[str, Any]=0.02 , snake_case :List[Any]=True , snake_case :Union[str, Any]=0.0 , snake_case :int=0.0 , snake_case :Union[str, Any]=None , snake_case :Dict=False , snake_case :int=False , snake_case :Tuple=True , snake_case :str=True , snake_case :List[Any]=False , snake_case :Optional[Any]=11 , snake_case :Tuple=11 , **snake_case :List[Any] , ): '''simple docstring''' A_ : Optional[int] = vocab_size # Backward compatibility with n_embed kwarg A_ : Any = kwargs.pop("n_embed" , snake_case ) A_ : str = hidden_size if n_embed is None else n_embed A_ : List[str] = num_hidden_layers A_ : List[str] = num_attention_heads A_ : List[str] = layer_norm_epsilon A_ : Optional[Any] = initializer_range A_ : Optional[int] = use_cache A_ : str = hidden_dropout A_ : str = attention_dropout A_ : str = bos_token_id A_ : List[str] = eos_token_id A_ : Union[str, Any] = num_attention_heads if num_kv_heads is None else num_kv_heads A_ : int = alibi A_ : str = new_decoder_architecture A_ : Dict = multi_query # Ignored when new_decoder_architecture is True A_ : Any = parallel_attn A_ : Optional[Any] = bias super().__init__(bos_token_id=snake_case , eos_token_id=snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' return self.hidden_size // self.num_attention_heads @property def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' return not self.alibi
358
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __snake_case ( ) -> tuple[list[int], int]: A_ : Dict = [randint(-1000 , 1000 ) for i in range(10 )] A_ : List[str] = randint(-5000 , 5000 ) return (arr, r) _lowerCAmelCase : List[Any] = make_dataset() def __snake_case ( _lowerCAmelCase : list[int] , _lowerCAmelCase : int ) -> tuple[int, ...]: for triplet in permutations(_lowerCAmelCase , 3 ): if sum(_lowerCAmelCase ) == target: return tuple(sorted(_lowerCAmelCase ) ) return (0, 0, 0) def __snake_case ( _lowerCAmelCase : list[int] , _lowerCAmelCase : int ) -> tuple[int, int, int]: arr.sort() A_ : Tuple = len(_lowerCAmelCase ) for i in range(n - 1 ): A_ , A_ : int = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __snake_case ( ) -> tuple[float, float]: A_ : Union[str, Any] = "\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n" A_ : Tuple = "\ntriplet_sum1(*dataset)\n" A_ : Optional[Any] = "\ntriplet_sum2(*dataset)\n" A_ : List[str] = repeat(setup=_lowerCAmelCase , stmt=_lowerCAmelCase , repeat=5 , number=10000 ) A_ : Tuple = repeat(setup=_lowerCAmelCase , stmt=_lowerCAmelCase , repeat=5 , number=10000 ) return (min(_lowerCAmelCase ), min(_lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _lowerCAmelCase : Optional[Any] = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
70
0
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class __UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): @register_to_config def __init__( self , lowerCAmelCase_ = 7_68 , ): """simple docstring""" super().__init__() _snake_case = nn.Parameter(torch.zeros(1 , lowerCAmelCase_ ) ) _snake_case = nn.Parameter(torch.ones(1 , lowerCAmelCase_ ) ) def lowerCamelCase ( self , lowerCAmelCase_ = None , lowerCAmelCase_ = None , ): """simple docstring""" _snake_case = nn.Parameter(self.mean.to(lowerCAmelCase_ ).to(lowerCAmelCase_ ) ) _snake_case = nn.Parameter(self.std.to(lowerCAmelCase_ ).to(lowerCAmelCase_ ) ) return self def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = (embeds * self.std) + self.mean return embeds
42
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
1
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _UpperCAmelCase : @staticmethod def a ( *_lowercase : List[Any] , **_lowercase : Union[str, Any] ): pass @is_pipeline_test @require_vision @require_timm @require_torch class _UpperCAmelCase ( unittest.TestCase ): a__ : Tuple = MODEL_FOR_OBJECT_DETECTION_MAPPING def a ( self : Tuple , _lowercase : Dict , _lowercase : Dict , _lowercase : Optional[Any] ): __UpperCAmelCase = ObjectDetectionPipeline(model=_lowercase , image_processor=_lowercase ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def a ( self : int , _lowercase : str , _lowercase : List[str] ): __UpperCAmelCase = object_detector('''./tests/fixtures/tests_samples/COCO/000000039769.png''' , threshold=0.0 ) self.assertGreater(len(_lowercase ) , 0 ) for detected_object in outputs: self.assertEqual( _lowercase , { '''score''': ANY(_lowercase ), '''label''': ANY(_lowercase ), '''box''': {'''xmin''': ANY(_lowercase ), '''ymin''': ANY(_lowercase ), '''xmax''': ANY(_lowercase ), '''ymax''': ANY(_lowercase )}, } , ) import datasets __UpperCAmelCase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) __UpperCAmelCase = [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] __UpperCAmelCase = object_detector(_lowercase , threshold=0.0 ) self.assertEqual(len(_lowercase ) , len(_lowercase ) ) for outputs in batch_outputs: self.assertGreater(len(_lowercase ) , 0 ) for detected_object in outputs: self.assertEqual( _lowercase , { '''score''': ANY(_lowercase ), '''label''': ANY(_lowercase ), '''box''': {'''xmin''': ANY(_lowercase ), '''ymin''': ANY(_lowercase ), '''xmax''': ANY(_lowercase ), '''ymax''': ANY(_lowercase )}, } , ) @require_tf @unittest.skip('''Object detection not implemented in TF''' ) def a ( self : int ): pass @require_torch def a ( self : Dict ): __UpperCAmelCase = '''hf-internal-testing/tiny-detr-mobilenetsv3''' __UpperCAmelCase = AutoModelForObjectDetection.from_pretrained(_lowercase ) __UpperCAmelCase = AutoFeatureExtractor.from_pretrained(_lowercase ) __UpperCAmelCase = ObjectDetectionPipeline(model=_lowercase , feature_extractor=_lowercase ) __UpperCAmelCase = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=0.0 ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, ] , ) __UpperCAmelCase = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ [ {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, ], [ {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, {'''score''': 0.3_376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 1_59, '''ymin''': 1_20, '''xmax''': 4_80, '''ymax''': 3_59}}, ], ] , ) @require_torch @slow def a ( self : List[str] ): __UpperCAmelCase = '''facebook/detr-resnet-50''' __UpperCAmelCase = AutoModelForObjectDetection.from_pretrained(_lowercase ) __UpperCAmelCase = AutoFeatureExtractor.from_pretrained(_lowercase ) __UpperCAmelCase = ObjectDetectionPipeline(model=_lowercase , feature_extractor=_lowercase ) __UpperCAmelCase = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ] , ) __UpperCAmelCase = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ], [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ], ] , ) @require_torch @slow def a ( self : Optional[Any] ): __UpperCAmelCase = '''facebook/detr-resnet-50''' __UpperCAmelCase = pipeline('''object-detection''' , model=_lowercase ) __UpperCAmelCase = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ] , ) __UpperCAmelCase = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ], [ {'''score''': 0.9_982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 1_75, '''ymax''': 1_17}}, {'''score''': 0.9_960, '''label''': '''remote''', '''box''': {'''xmin''': 3_33, '''ymin''': 72, '''xmax''': 3_68, '''ymax''': 1_87}}, {'''score''': 0.9_955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 6_39, '''ymax''': 4_73}}, {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ], ] , ) @require_torch @slow def a ( self : Any ): __UpperCAmelCase = 0.9_985 __UpperCAmelCase = '''facebook/detr-resnet-50''' __UpperCAmelCase = pipeline('''object-detection''' , model=_lowercase ) __UpperCAmelCase = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=_lowercase ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ {'''score''': 0.9_988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 3_14, '''ymax''': 4_70}}, {'''score''': 0.9_987, '''label''': '''cat''', '''box''': {'''xmin''': 3_45, '''ymin''': 23, '''xmax''': 6_40, '''ymax''': 3_68}}, ] , ) @require_torch @require_pytesseract @slow def a ( self : Any ): __UpperCAmelCase = '''Narsil/layoutlmv3-finetuned-funsd''' __UpperCAmelCase = 0.9_993 __UpperCAmelCase = pipeline('''object-detection''' , model=_lowercase , threshold=_lowercase ) __UpperCAmelCase = object_detector( '''https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png''' ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ {'''score''': 0.9_993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 2_94, '''ymin''': 2_54, '''xmax''': 3_43, '''ymax''': 2_64}}, {'''score''': 0.9_993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 2_94, '''ymin''': 2_54, '''xmax''': 3_43, '''ymax''': 2_64}}, ] , )
86
"""simple docstring""" from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 _lowercase : str = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 2048-bit 14: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 3072-bit 15: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 4096-bit 16: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 6144-bit 17: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 8192-bit 18: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, } class _UpperCAmelCase : def __init__( self : List[Any] , _lowercase : int = 14 ): if group not in primes: raise ValueError('''Unsupported Group''' ) __UpperCAmelCase = primes[group]['''prime'''] __UpperCAmelCase = primes[group]['''generator'''] __UpperCAmelCase = int(hexlify(urandom(32 ) ) , base=16 ) def a ( self : int ): return hex(self.__private_key )[2:] def a ( self : Dict ): __UpperCAmelCase = pow(self.generator , self.__private_key , self.prime ) return hex(_lowercase )[2:] def a ( self : Union[str, Any] , _lowercase : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(_lowercase , (self.prime - 1) // 2 , self.prime ) == 1 ) def a ( self : Optional[Any] , _lowercase : str ): __UpperCAmelCase = int(_lowercase , base=16 ) if not self.is_valid_public_key(_lowercase ): raise ValueError('''Invalid public key''' ) __UpperCAmelCase = pow(_lowercase , self.__private_key , self.prime ) return shaaaa(str(_lowercase ).encode() ).hexdigest() @staticmethod def a ( _lowercase : int , _lowercase : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(_lowercase , (prime - 1) // 2 , _lowercase ) == 1 ) @staticmethod def a ( _lowercase : str , _lowercase : str , _lowercase : int = 14 ): __UpperCAmelCase = int(_lowercase , base=16 ) __UpperCAmelCase = int(_lowercase , base=16 ) __UpperCAmelCase = primes[group]['''prime'''] if not DiffieHellman.is_valid_public_key_static(_lowercase , _lowercase ): raise ValueError('''Invalid public key''' ) __UpperCAmelCase = pow(_lowercase , _lowercase , _lowercase ) return shaaaa(str(_lowercase ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
86
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=_lowercase) class __snake_case ( _lowercase): snake_case__ : str = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True}) snake_case__ : ClassVar[Features] = Features({"image": Image()}) snake_case__ : ClassVar[Features] = Features({"labels": ClassLabel}) snake_case__ : str = "image" snake_case__ : str = "labels" def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : Dict ): """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __lowerCAmelCase ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) _lowerCamelCase : List[Any] = copy.deepcopy(self ) _lowerCamelCase : str = self.label_schema.copy() _lowerCamelCase : Union[str, Any] = features[self.label_column] _lowerCamelCase : List[str] = label_schema return task_template @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
72
import torch def UpperCAmelCase__ ( ): if torch.cuda.is_available(): lowercase :Optional[int] = torch.cuda.device_count() else: lowercase :Dict = 0 print(F"Successfully ran on {num_gpus} GPUs" ) if __name__ == "__main__": main()
236
0
"""simple docstring""" import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { "facebook/encodec_24khz": "https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json", "facebook/encodec_48khz": "https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json", } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'encodec' def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=[1.5, 3.0, 6.0, 12.0, 24.0] , _SCREAMING_SNAKE_CASE: List[str]=2_4000 , _SCREAMING_SNAKE_CASE: List[str]=1 , _SCREAMING_SNAKE_CASE: int=False , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Any=128 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Any=[8, 5, 4, 2] , _SCREAMING_SNAKE_CASE: List[Any]="weight_norm" , _SCREAMING_SNAKE_CASE: Optional[int]=7 , _SCREAMING_SNAKE_CASE: Optional[int]=7 , _SCREAMING_SNAKE_CASE: Optional[int]=3 , _SCREAMING_SNAKE_CASE: Tuple=2 , _SCREAMING_SNAKE_CASE: Tuple=True , _SCREAMING_SNAKE_CASE: Tuple="reflect" , _SCREAMING_SNAKE_CASE: Optional[Any]=2 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: Any=1.0 , _SCREAMING_SNAKE_CASE: int=1024 , _SCREAMING_SNAKE_CASE: Tuple=None , _SCREAMING_SNAKE_CASE: int=True , **_SCREAMING_SNAKE_CASE: Optional[Any] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = target_bandwidths __lowerCAmelCase : Union[str, Any] = sampling_rate __lowerCAmelCase : Tuple = audio_channels __lowerCAmelCase : Any = normalize __lowerCAmelCase : Any = chunk_length_s __lowerCAmelCase : Dict = overlap __lowerCAmelCase : Any = hidden_size __lowerCAmelCase : Dict = num_filters __lowerCAmelCase : Optional[Any] = num_residual_layers __lowerCAmelCase : Any = upsampling_ratios __lowerCAmelCase : Dict = norm_type __lowerCAmelCase : Tuple = kernel_size __lowerCAmelCase : str = last_kernel_size __lowerCAmelCase : Any = residual_kernel_size __lowerCAmelCase : List[Any] = dilation_growth_rate __lowerCAmelCase : str = use_causal_conv __lowerCAmelCase : Tuple = pad_mode __lowerCAmelCase : Any = compress __lowerCAmelCase : str = num_lstm_layers __lowerCAmelCase : List[str] = trim_right_ratio __lowerCAmelCase : Tuple = codebook_size __lowerCAmelCase : List[Any] = codebook_dim if codebook_dim is not None else hidden_size __lowerCAmelCase : Optional[int] = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F"""self.norm_type must be one of `\"weight_norm\"`, `\"time_group_norm\"`), got {self.norm_type}""") super().__init__(**_a) @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[Any]: """simple docstring""" if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate) @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length)) @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = np.prod(self.upsampling_ratios) return math.ceil(self.sampling_rate / hop_length) @property def _SCREAMING_SNAKE_CASE ( self: int) -> int: """simple docstring""" return int(1000 * self.target_bandwidths[-1] // (self.frame_rate * 10))
369
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : List[str] = { 'configuration_timesformer': ['TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimesformerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimesformerModel', 'TimesformerForVideoClassification', 'TimesformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
58
0
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __A = logging.get_logger(__name__) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :int = ["pixel_values"] def __init__( self , _UpperCAmelCase = True , _UpperCAmelCase = 1 / 255 , _UpperCAmelCase = True , _UpperCAmelCase = 8 , **_UpperCAmelCase , ): super().__init__(**_UpperCAmelCase ) lowercase__: Optional[Any] = do_rescale lowercase__: Tuple = rescale_factor lowercase__: int = do_pad lowercase__: Union[str, Any] = pad_size def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , **_UpperCAmelCase ): return rescale(_UpperCAmelCase , scale=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None ): lowercase__, lowercase__: int = get_image_size(_UpperCAmelCase ) lowercase__: int = (old_height // size + 1) * size - old_height lowercase__: Dict = (old_width // size + 1) * size - old_width return pad(_UpperCAmelCase , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=_UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = ChannelDimension.FIRST , **_UpperCAmelCase , ): lowercase__: Any = do_rescale if do_rescale is not None else self.do_rescale lowercase__: Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__: int = do_pad if do_pad is not None else self.do_pad lowercase__: int = pad_size if pad_size is not None else self.pad_size lowercase__: List[str] = make_list_of_images(_UpperCAmelCase ) if not valid_images(_UpperCAmelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. lowercase__: Optional[Any] = [to_numpy_array(_UpperCAmelCase ) for image in images] if do_rescale: lowercase__: int = [self.rescale(image=_UpperCAmelCase , scale=_UpperCAmelCase ) for image in images] if do_pad: lowercase__: Optional[Any] = [self.pad(_UpperCAmelCase , size=_UpperCAmelCase ) for image in images] lowercase__: List[Any] = [to_channel_dimension_format(_UpperCAmelCase , _UpperCAmelCase ) for image in images] lowercase__: List[Any] = {'''pixel_values''': images} return BatchFeature(data=_UpperCAmelCase , tensor_type=_UpperCAmelCase )
177
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class UpperCAmelCase (unittest.TestCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase=13 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=99 , _UpperCAmelCase=32 , _UpperCAmelCase=5 , _UpperCAmelCase=4 , _UpperCAmelCase=37 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=16 , _UpperCAmelCase=2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=4 , ): lowercase__: Dict = parent lowercase__: List[str] = batch_size lowercase__: Optional[Any] = seq_length lowercase__: List[Any] = is_training lowercase__: int = use_attention_mask lowercase__: Tuple = use_token_type_ids lowercase__: Union[str, Any] = use_labels lowercase__: str = vocab_size lowercase__: str = hidden_size lowercase__: str = num_hidden_layers lowercase__: Optional[int] = num_attention_heads lowercase__: List[str] = intermediate_size lowercase__: List[str] = hidden_act lowercase__: Tuple = hidden_dropout_prob lowercase__: int = attention_probs_dropout_prob lowercase__: int = max_position_embeddings lowercase__: Union[str, Any] = type_vocab_size lowercase__: List[Any] = type_sequence_label_size lowercase__: Any = initializer_range lowercase__: str = num_choices def _snake_case ( self ): lowercase__: int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__: List[Any] = None if self.use_attention_mask: lowercase__: Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__: List[Any] = None if self.use_token_type_ids: lowercase__: str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__: Optional[int] = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_UpperCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _snake_case ( self ): lowercase__: str = self.prepare_config_and_inputs() lowercase__, lowercase__, lowercase__, lowercase__: Optional[Any] = config_and_inputs lowercase__: Dict = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" _UpperCAmelCase :List[str] = True _UpperCAmelCase :Union[str, Any] = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def _snake_case ( self ): lowercase__: str = FlaxRoFormerModelTester(self ) @slow def _snake_case ( self ): for model_class_name in self.all_model_classes: lowercase__: Dict = model_class_name.from_pretrained('''junnyu/roformer_chinese_small''' , from_pt=_UpperCAmelCase ) lowercase__: int = model(np.ones((1, 1) ) ) self.assertIsNotNone(_UpperCAmelCase ) @require_flax class UpperCAmelCase (unittest.TestCase ): """simple docstring""" @slow def _snake_case ( self ): lowercase__: Any = FlaxRoFormerForMaskedLM.from_pretrained('''junnyu/roformer_chinese_base''' ) lowercase__: Optional[int] = jnp.array([[0, 1, 2, 3, 4, 5]] ) lowercase__: List[Any] = model(_UpperCAmelCase )[0] lowercase__: str = 50000 lowercase__: Tuple = (1, 6, vocab_size) self.assertEqual(output.shape , _UpperCAmelCase ) lowercase__: List[Any] = jnp.array( [[[-0.1_205, -1.0_265, 0.2_922], [-1.5_134, 0.1_974, 0.1_519], [-5.0_135, -3.9_003, -0.8_404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 ) )
177
1
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return int((input_a, input_a).count(0 ) != 0 ) def a__ ( ) -> None: """simple docstring""" assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
67
'''simple docstring''' from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder _lowerCamelCase = datasets.utils.logging.get_logger(__name__) class _snake_case (folder_based_builder.FolderBasedBuilderConfig): __A : bool =None __A : bool =None class _snake_case (folder_based_builder.FolderBasedBuilder): __A : Union[str, Any] =datasets.Audio() __A : Optional[int] ="audio" __A : Any =AudioFolderConfig __A : List[str] # definition at the bottom of the script __A : Optional[int] =AudioClassification(audio_column="audio" , label_column="label") _lowerCamelCase = [ """.aiff""", """.au""", """.avr""", """.caf""", """.flac""", """.htk""", """.svx""", """.mat4""", """.mat5""", """.mpc2k""", """.ogg""", """.paf""", """.pvf""", """.raw""", """.rf64""", """.sd2""", """.sds""", """.ircam""", """.voc""", """.w64""", """.wav""", """.nist""", """.wavex""", """.wve""", """.xi""", """.mp3""", """.opus""", ] _lowerCamelCase = AUDIO_EXTENSIONS
67
1
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers A_ :Dict = [ '''python''', '''tqdm''', '''regex''', '''requests''', '''packaging''', '''filelock''', '''numpy''', '''tokenizers''', '''huggingface-hub''', '''safetensors''', '''accelerate''', '''pyyaml''', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"can't find {pkg} in {deps.keys()}, check dependency_versions_table.py") def A ( a_ ,a_=None ) -> Union[str, Any]: require_version(deps[pkg] ,a_ )
71
from collections import defaultdict def _snake_case ( lowerCAmelCase : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = 1 SCREAMING_SNAKE_CASE_ : Tuple = True for v in tree[start]: if v not in visited: ret += dfs(lowerCAmelCase ) if ret % 2 == 0: cuts.append(lowerCAmelCase ) return ret def _snake_case ( ): """simple docstring""" dfs(1 ) if __name__ == "__main__": __lowerCamelCase , __lowerCamelCase : Union[str, Any] = 10, 9 __lowerCamelCase : Optional[int] = defaultdict(list) __lowerCamelCase : dict[int, bool] = {} __lowerCamelCase : list[int] = [] __lowerCamelCase : Optional[Any] = 0 __lowerCamelCase : Any = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
18
0
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset _SCREAMING_SNAKE_CASE : Union[str, Any] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class UpperCAmelCase__ ( nn.Module ): """simple docstring""" def __init__( self : int , __lowerCamelCase : str ) -> Any: super().__init__() SCREAMING_SNAKE_CASE__ = torchvision.models.resnetaaa(pretrained=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = list(model.children() )[:-2] SCREAMING_SNAKE_CASE__ = nn.Sequential(*__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def lowercase_ ( self : str , __lowerCamelCase : str ) -> List[Any]: # Bx3x224x224 -> Bx2048x7x7 -> Bx2048xN -> BxNx2048 SCREAMING_SNAKE_CASE__ = self.pool(self.model(__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__ = torch.flatten(__lowerCamelCase , start_dim=2 ) SCREAMING_SNAKE_CASE__ = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class UpperCAmelCase__ ( A__ ): """simple docstring""" def __init__( self : Any , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : str ) -> List[str]: SCREAMING_SNAKE_CASE__ = [json.loads(__lowerCamelCase ) for l in open(__lowerCamelCase )] SCREAMING_SNAKE_CASE__ = os.path.dirname(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = tokenizer SCREAMING_SNAKE_CASE__ = labels SCREAMING_SNAKE_CASE__ = len(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = max_seq_length SCREAMING_SNAKE_CASE__ = transforms def __len__( self : Dict ) -> str: return len(self.data ) def __getitem__( self : Any , __lowerCamelCase : Dict ) -> Tuple: SCREAMING_SNAKE_CASE__ = torch.LongTensor(self.tokenizer.encode(self.data[index]['''text'''] , add_special_tokens=__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = sentence[0], sentence[1:-1], sentence[-1] SCREAMING_SNAKE_CASE__ = sentence[: self.max_seq_length] SCREAMING_SNAKE_CASE__ = torch.zeros(self.n_classes ) SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = Image.open(os.path.join(self.data_dir , self.data[index]['''img'''] ) ).convert('''RGB''' ) SCREAMING_SNAKE_CASE__ = self.transforms(__lowerCamelCase ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def lowercase_ ( self : Any ) -> Tuple: SCREAMING_SNAKE_CASE__ = Counter() for row in self.data: label_freqs.update(row['''label'''] ) return label_freqs def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = [len(row['''sentence'''] ) for row in batch] SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = len(_A ), max(_A ) SCREAMING_SNAKE_CASE__ = torch.zeros(_A , _A , dtype=torch.long ) SCREAMING_SNAKE_CASE__ = torch.zeros(_A , _A , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(_A , _A ) ): SCREAMING_SNAKE_CASE__ = input_row['''sentence'''] SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.stack([row['''image'''] for row in batch] ) SCREAMING_SNAKE_CASE__ = torch.stack([row['''label'''] for row in batch] ) SCREAMING_SNAKE_CASE__ = torch.stack([row['''image_start_token'''] for row in batch] ) SCREAMING_SNAKE_CASE__ = torch.stack([row['''image_end_token'''] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def UpperCAmelCase_ ( ): '''simple docstring''' return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def UpperCAmelCase_ ( ): '''simple docstring''' return transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize( mean=[0.4_6_7_7_7_0_4_4, 0.4_4_5_3_1_4_2_9, 0.4_0_6_6_1_0_1_7] , std=[0.1_2_2_2_1_9_9_4, 0.1_2_1_4_5_8_3_5, 0.1_4_3_8_0_4_6_9] , ), ] )
218
def UpperCAmelCase_ ( _A ): '''simple docstring''' if not isinstance(_A , _A ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''multiplicative_persistence() does not accept negative values''' ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = str(_A ) while len(_A ) != 1: SCREAMING_SNAKE_CASE__ = [int(_A ) for i in num_string] SCREAMING_SNAKE_CASE__ = 1 for i in range(0 , len(_A ) ): total *= numbers[i] SCREAMING_SNAKE_CASE__ = str(_A ) steps += 1 return steps def UpperCAmelCase_ ( _A ): '''simple docstring''' if not isinstance(_A , _A ): raise ValueError('''additive_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''additive_persistence() does not accept negative values''' ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = str(_A ) while len(_A ) != 1: SCREAMING_SNAKE_CASE__ = [int(_A ) for i in num_string] SCREAMING_SNAKE_CASE__ = 0 for i in range(0 , len(_A ) ): total += numbers[i] SCREAMING_SNAKE_CASE__ = str(_A ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
218
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: lowercase_ = None lowercase_ = logging.get_logger(__name__) lowercase_ = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} lowercase_ = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", }, "tokenizer_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json", }, } lowercase_ = { "albert-base-v1": 5_1_2, "albert-large-v1": 5_1_2, "albert-xlarge-v1": 5_1_2, "albert-xxlarge-v1": 5_1_2, "albert-base-v2": 5_1_2, "albert-large-v2": 5_1_2, "albert-xlarge-v2": 5_1_2, "albert-xxlarge-v2": 5_1_2, } lowercase_ = "▁" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : List[str] = VOCAB_FILES_NAMES __UpperCAmelCase : List[Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : int = AlbertTokenizer def __init__( self , _a=None , _a=None , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , **_a , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __a = ( AddedToken(_a , lstrip=_a , rstrip=_a , normalized=_a ) if isinstance(_a , _a ) else mask_token ) super().__init__( _a , tokenizer_file=_a , do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , **_a , ) __a = do_lower_case __a = remove_space __a = keep_accents __a = vocab_file __a = False if not self.vocab_file else True def __UpperCAmelCase ( self , _a , _a = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __UpperCAmelCase ( self , _a , _a = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _a , _a = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __a = os.path.join( _a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) return (out_vocab_file,)
45
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def lowercase ( lowerCAmelCase__ : Namespace ) -> Tuple: return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) lowercase_ = "\ntransformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires\nTensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions.\n" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @staticmethod def __UpperCAmelCase ( _a ): __a = parser.add_parser( '''convert''' , help='''CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints.''' , ) train_parser.add_argument('''--model_type''' , type=_a , required=_a , help='''Model\'s type.''' ) train_parser.add_argument( '''--tf_checkpoint''' , type=_a , required=_a , help='''TensorFlow checkpoint path or folder.''' ) train_parser.add_argument( '''--pytorch_dump_output''' , type=_a , required=_a , help='''Path to the PyTorch saved model output.''' ) train_parser.add_argument('''--config''' , type=_a , default='''''' , help='''Configuration file path or folder.''' ) train_parser.add_argument( '''--finetuning_task_name''' , type=_a , default=_a , help='''Optional fine-tuning task name if the TF model was a finetuned model.''' , ) train_parser.set_defaults(func=_a ) def __init__( self , _a , _a , _a , _a , _a , *_a , ): __a = logging.get_logger('''transformers-cli/converting''' ) self._logger.info(f'''Loading model {model_type}''' ) __a = model_type __a = tf_checkpoint __a = pytorch_dump_output __a = config __a = finetuning_task_name def __UpperCAmelCase ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(_a ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) if "ckpt" in self._tf_checkpoint.lower(): __a = self._tf_checkpoint __a = '''''' else: __a = self._tf_checkpoint __a = '''''' convert_transfo_xl_checkpoint_to_pytorch( _a , self._config , self._pytorch_dump_output , _a ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_a ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( '''--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]''' )
45
1
"""simple docstring""" def __lowerCamelCase ( a_ : str , a_ : int ) -> Any: __SCREAMING_SNAKE_CASE :Optional[Any] = word.split() def justify(a_ : list , a_ : int , a_ : int ) -> str: __SCREAMING_SNAKE_CASE :Dict = max_width - width __SCREAMING_SNAKE_CASE :str = len(UpperCAmelCase_ ) if len(UpperCAmelCase_ ) == 1: # if there is only word in line # just insert overall_spaces_count for the remainder of line return line[0] + " " * overall_spaces_count else: __SCREAMING_SNAKE_CASE :int = words_count - 1 # num_spaces_between_words_list[i] : tells you to insert # num_spaces_between_words_list[i] spaces # after word on line[i] __SCREAMING_SNAKE_CASE :Tuple = spaces_to_insert_between_words * [ overall_spaces_count // spaces_to_insert_between_words ] __SCREAMING_SNAKE_CASE :Optional[int] = ( overall_spaces_count % spaces_to_insert_between_words ) # distribute spaces via round robin to the left words for i in range(UpperCAmelCase_ ): num_spaces_between_words_list[i] += 1 __SCREAMING_SNAKE_CASE :Optional[Any] = [] for i in range(UpperCAmelCase_ ): # add the word aligned_words_list.append(line[i] ) # add the spaces to insert aligned_words_list.append(num_spaces_between_words_list[i] * ''' ''' ) # just add the last word to the sentence aligned_words_list.append(line[-1] ) # join the aligned words list to form a justified line return "".join(UpperCAmelCase_ ) __SCREAMING_SNAKE_CASE :Optional[int] = [] __SCREAMING_SNAKE_CASE :Optional[Any] = [] __SCREAMING_SNAKE_CASE :List[Any] = 0 for word in words: if width + len(UpperCAmelCase_ ) + len(UpperCAmelCase_ ) <= max_width: # keep adding words until we can fill out max_width # width = sum of length of all words (without overall_spaces_count) # len(word) = length of current word # len(line) = number of overall_spaces_count to insert between words line.append(UpperCAmelCase_ ) width += len(UpperCAmelCase_ ) else: # justify the line and add it to result answer.append(justify(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) ) # reset new line and new width __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :List[str] = [word], len(UpperCAmelCase_ ) __SCREAMING_SNAKE_CASE :Tuple = max_width - width - len(UpperCAmelCase_ ) answer.append(''' '''.join(UpperCAmelCase_ ) + (remaining_spaces + 1) * ''' ''' ) return answer if __name__ == "__main__": from doctest import testmod testmod()
353
"""simple docstring""" def __lowerCamelCase ( a_ : str , a_ : str ) -> str: __SCREAMING_SNAKE_CASE :int = len(a_ ) __SCREAMING_SNAKE_CASE :int = len(a_ ) __SCREAMING_SNAKE_CASE :int = ( first_str_length if first_str_length > second_str_length else second_str_length ) __SCREAMING_SNAKE_CASE :list = [] for char_count in range(a_ ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(a_ ) if __name__ == "__main__": print(alternative_string_arrange("AB", "XYZ"), end=" ")
239
0
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Optional[Any] = MODEL_FOR_MASKED_LM_MAPPING lowerCamelCase_ : List[Any] = TF_MODEL_FOR_MASKED_LM_MAPPING def _lowercase ( self ) -> List[Any]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def _lowercase ( self ) -> List[str]: lowerCamelCase : str = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf" ) lowerCamelCase : Dict = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ {"sequence": "My name is grouped", "score": 2.1e-05, "token": 3_8015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-05, "token": 2_5506, "token_str": " accuser"}, ] , ) lowerCamelCase : int = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-05, "token": 3_8015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-05, "token": 2_5506, "token_str": " accuser", }, ] , ) lowerCamelCase : List[str] = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Clara", "score": 2e-05, "token": 1_3606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-05, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def _lowercase ( self ) -> List[str]: lowerCamelCase : List[Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt" ) lowerCamelCase : Tuple = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Maul", "score": 2.2e-05, "token": 3_5676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-05, "token": 1_6416, "token_str": "ELS"}, ] , ) lowerCamelCase : Union[str, Any] = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-05, "token": 3_5676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-05, "token": 1_6416, "token_str": "ELS"}, ] , ) lowerCamelCase : int = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Patrick", "score": 2.1e-05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-05, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-05, "token": 1_3606, "token_str": " Clara"}, ] , ) lowerCamelCase : int = unmasker("My name is <mask> <mask>" , top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ , decimals=6 ) , [ [ { "score": 2.2e-05, "token": 3_5676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-05, "token": 1_6416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-05, "token": 3_5676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-05, "token": 1_6416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def _lowercase ( self ) -> Dict: lowerCamelCase : Any = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt" ) # convert model to fp16 pipe.model.half() lowerCamelCase : Tuple = pipe("Paris is the [MASK] of France." ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) @slow @require_torch def _lowercase ( self ) -> List[Any]: lowerCamelCase : Tuple = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt" ) self.run_large_test(UpperCamelCase__ ) @slow @require_tf def _lowercase ( self ) -> str: lowerCamelCase : Tuple = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf" ) self.run_large_test(UpperCamelCase__ ) def _lowercase ( self , UpperCamelCase__ ) -> Optional[int]: lowerCamelCase : List[Any] = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ {"sequence": "My name is John", "score": 0.008, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.007, "token": 1573, "token_str": " Chris"}, ] , ) lowerCamelCase : List[Any] = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ { "sequence": "The largest city in France is Paris", "score": 0.251, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.214, "token": 1_2790, "token_str": " Lyon", }, ] , ) lowerCamelCase : int = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ {"sequence": "My name is Patrick", "score": 0.005, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.000, "token": 1_3606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.000, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def _lowercase ( self ) -> List[str]: lowerCamelCase : Union[str, Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt" ) lowerCamelCase : List[str] = None lowerCamelCase : Dict = None self.run_pipeline_test(UpperCamelCase__ , [] ) @require_tf def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : Optional[int] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf" ) lowerCamelCase : int = None lowerCamelCase : List[Any] = None self.run_pipeline_test(UpperCamelCase__ , [] ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)" ) lowerCamelCase : Union[str, Any] = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = [ F'''This is another {tokenizer.mask_token} test''', ] return fill_masker, examples def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> int: lowerCamelCase : Tuple = fill_masker.tokenizer lowerCamelCase : Optional[Any] = fill_masker.model lowerCamelCase : Optional[int] = fill_masker( F'''This is a {tokenizer.mask_token}''' , ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) lowerCamelCase : int = fill_masker([F'''This is a {tokenizer.mask_token}'''] ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) lowerCamelCase : Any = fill_masker([F'''This is a {tokenizer.mask_token}''', F'''Another {tokenizer.mask_token} great test.'''] ) self.assertEqual( UpperCamelCase__ , [ [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ], [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ], ] , ) with self.assertRaises(UpperCamelCase__ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(UpperCamelCase__ ): fill_masker("This is" ) self.run_test_top_k(UpperCamelCase__ , UpperCamelCase__ ) self.run_test_targets(UpperCamelCase__ , UpperCamelCase__ ) self.run_test_top_k_targets(UpperCamelCase__ , UpperCamelCase__ ) self.fill_mask_with_duplicate_targets_and_top_k(UpperCamelCase__ , UpperCamelCase__ ) self.fill_mask_with_multiple_masks(UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Any: lowerCamelCase : Dict = tokenizer.get_vocab() lowerCamelCase : Tuple = sorted(vocab.keys() )[:2] # Pipeline argument lowerCamelCase : int = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ , targets=UpperCamelCase__ ) lowerCamelCase : List[Any] = fill_masker(F'''This is a {tokenizer.mask_token}''' ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) lowerCamelCase : List[str] = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , UpperCamelCase__ ) lowerCamelCase : Any = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(UpperCamelCase__ ) ) # Call argument lowerCamelCase : Tuple = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) lowerCamelCase : Tuple = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) lowerCamelCase : List[str] = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , UpperCamelCase__ ) lowerCamelCase : List[str] = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(UpperCamelCase__ ) ) # Score equivalence lowerCamelCase : Optional[Any] = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = [top_mask["token_str"] for top_mask in outputs] lowerCamelCase : Union[str, Any] = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(UpperCamelCase__ ) == set(UpperCamelCase__ ): lowerCamelCase : Any = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets=UpperCamelCase__ ) lowerCamelCase : List[str] = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(UpperCamelCase__ ) , nested_simplify(UpperCamelCase__ ) ) # Raises with invalid with self.assertRaises(UpperCamelCase__ ): lowerCamelCase : Any = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(UpperCamelCase__ ): lowerCamelCase : int = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets=[""] ) with self.assertRaises(UpperCamelCase__ ): lowerCamelCase : int = fill_masker(F'''This is a {tokenizer.mask_token}''' , targets="" ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : Union[str, Any] = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ , top_k=2 ) lowerCamelCase : Union[str, Any] = fill_masker(F'''This is a {tokenizer.mask_token}''' ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) lowerCamelCase : Union[str, Any] = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) lowerCamelCase : int = fill_masker(F'''This is a {tokenizer.mask_token}''' , top_k=2 ) self.assertEqual( UpperCamelCase__ , [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ] , ) self.assertEqual(nested_simplify(UpperCamelCase__ ) , nested_simplify(UpperCamelCase__ ) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> int: lowerCamelCase : Tuple = tokenizer.get_vocab() lowerCamelCase : Any = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) # top_k=2, ntargets=3 lowerCamelCase : Any = sorted(vocab.keys() )[:3] lowerCamelCase : Dict = fill_masker(F'''This is a {tokenizer.mask_token}''' , top_k=2 , targets=UpperCamelCase__ ) # If we use the most probably targets, and filter differently, we should still # have the same results lowerCamelCase : List[Any] = [el["token_str"] for el in sorted(UpperCamelCase__ , key=lambda UpperCamelCase__ : x["score"] , reverse=UpperCamelCase__ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(UpperCamelCase__ ).issubset(UpperCamelCase__ ): lowerCamelCase : int = fill_masker(F'''This is a {tokenizer.mask_token}''' , top_k=3 , targets=UpperCamelCase__ ) # They should yield exactly the same result self.assertEqual(nested_simplify(UpperCamelCase__ ) , nested_simplify(UpperCamelCase__ ) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> str: lowerCamelCase : int = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) lowerCamelCase : str = tokenizer.get_vocab() # String duplicates + id duplicates lowerCamelCase : Tuple = sorted(vocab.keys() )[:3] lowerCamelCase : List[str] = [targets[0], targets[1], targets[0], targets[2], targets[1]] lowerCamelCase : int = fill_masker(F'''My name is {tokenizer.mask_token}''' , targets=UpperCamelCase__ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(UpperCamelCase__ ) , 3 ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: lowerCamelCase : int = FillMaskPipeline(model=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) lowerCamelCase : Tuple = fill_masker( F'''This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}''' , top_k=2 ) self.assertEqual( UpperCamelCase__ , [ [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ], [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ], [ {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, {"sequence": ANY(UpperCamelCase__ ), "score": ANY(UpperCamelCase__ ), "token": ANY(UpperCamelCase__ ), "token_str": ANY(UpperCamelCase__ )}, ], ] , )
48
import math def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> float: if ( not isinstance(_SCREAMING_SNAKE_CASE ,(int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * power_factor def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> float: if ( not isinstance(_SCREAMING_SNAKE_CASE ,(int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
48
1
'''simple docstring''' def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return number | (1 << position) def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return number & ~(1 << position) def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return number ^ (1 << position) def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> bool: """simple docstring""" return ((number >> position) & 1) == 1 def _snake_case ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
187
'''simple docstring''' from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class __snake_case: '''simple docstring''' def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.0_2 , A_=3 , A_=4 , A_=None , ) -> Dict: lowerCAmelCase = parent lowerCAmelCase = 13 lowerCAmelCase = 7 lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = 99 lowerCAmelCase = 384 lowerCAmelCase = 2 lowerCAmelCase = 4 lowerCAmelCase = 37 lowerCAmelCase = """gelu""" lowerCAmelCase = 0.1 lowerCAmelCase = 0.1 lowerCAmelCase = 512 lowerCAmelCase = 16 lowerCAmelCase = 2 lowerCAmelCase = 0.0_2 lowerCAmelCase = 3 lowerCAmelCase = 4 lowerCAmelCase = 128 lowerCAmelCase = 2 lowerCAmelCase = 9 lowerCAmelCase = 1 lowerCAmelCase = None def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = None if self.use_input_mask: lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> int: lowerCAmelCase = TFConvBertModel(config=A_ ) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = [input_ids, input_mask] lowerCAmelCase = model(A_ ) lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> List[Any]: lowerCAmelCase = TFConvBertForMaskedLM(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Optional[int]: lowerCAmelCase = self.num_labels lowerCAmelCase = TFConvBertForSequenceClassification(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Any: lowerCAmelCase = self.num_choices lowerCAmelCase = TFConvBertForMultipleChoice(config=A_ ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Union[str, Any]: lowerCAmelCase = self.num_labels lowerCAmelCase = TFConvBertForTokenClassification(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Optional[int]: lowerCAmelCase = TFConvBertForQuestionAnswering(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __snake_case ( self ) -> Any: lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ) = config_and_inputs lowerCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __snake_case( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Optional[int] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase : Union[str, Any] = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase : Union[str, Any] = False UpperCAmelCase : Optional[int] = False UpperCAmelCase : Dict = False def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = TFConvBertModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __snake_case ( self ) -> Tuple: self.config_tester.run_common_tests() def __snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __snake_case ( self ) -> List[str]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __snake_case ( self ) -> str: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __snake_case ( self ) -> Any: lowerCAmelCase, lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = True lowerCAmelCase = True if hasattr(A_ , """use_cache""" ): lowerCAmelCase = True lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) for model_class in self.all_model_classes: lowerCAmelCase = self._prepare_for_class(A_ , A_ ) lowerCAmelCase = model_class(A_ ) lowerCAmelCase = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) lowerCAmelCase = os.path.join(A_ , """saved_model""" , """1""" ) lowerCAmelCase = tf.keras.models.load_model(A_ ) lowerCAmelCase = model(A_ ) if self.is_encoder_decoder: lowerCAmelCase = outputs["""encoder_hidden_states"""] lowerCAmelCase = outputs["""encoder_attentions"""] else: lowerCAmelCase = outputs["""hidden_states"""] lowerCAmelCase = outputs["""attentions"""] self.assertEqual(len(A_ ) , A_ ) lowerCAmelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __snake_case ( self ) -> Optional[Any]: lowerCAmelCase = TFConvBertModel.from_pretrained("""YituTech/conv-bert-base""" ) self.assertIsNotNone(A_ ) def __snake_case ( self ) -> str: lowerCAmelCase, lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = True lowerCAmelCase = getattr(self.model_tester , """decoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) def check_decoder_attentions_output(A_ ): lowerCAmelCase = len(A_ ) self.assertEqual(out_len % 2 , 0 ) lowerCAmelCase = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): lowerCAmelCase = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: lowerCAmelCase = True lowerCAmelCase = False lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) lowerCAmelCase = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowerCAmelCase = True lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class __snake_case( unittest.TestCase ): '''simple docstring''' @slow def __snake_case ( self ) -> Any: lowerCAmelCase = TFConvBertModel.from_pretrained("""YituTech/conv-bert-base""" ) lowerCAmelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase = model(A_ )[0] lowerCAmelCase = [1, 6, 768] self.assertEqual(output.shape , A_ ) lowerCAmelCase = tf.constant( [ [ [-0.0_3_4_7_5_4_9_3, -0.4_6_8_6_0_3_4, -0.3_0_6_3_8_8_3_2], [0.2_2_6_3_7_2_4_8, -0.2_6_9_8_8_6_4_6, -0.7_4_2_3_4_2_4], [0.1_0_3_2_4_8_6_8, -0.4_5_0_1_3_5_0_8, -0.5_8_2_8_0_7_8_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
187
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , A_ , A_=7 , A_=3 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=True , A_=1 / 255 , A_=True , ) -> Optional[int]: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __UpperCamelCase =size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} __UpperCamelCase =parent __UpperCamelCase =batch_size __UpperCamelCase =num_channels __UpperCamelCase =min_resolution __UpperCamelCase =max_resolution __UpperCamelCase =do_resize __UpperCamelCase =size __UpperCamelCase =do_normalize __UpperCamelCase =image_mean __UpperCamelCase =image_std __UpperCamelCase =do_rescale __UpperCamelCase =rescale_factor __UpperCamelCase =do_pad def _a ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _a ( self , A_ , A_=False ) -> Optional[int]: if not batched: __UpperCamelCase =image_inputs[0] if isinstance(A_ , Image.Image ): __UpperCamelCase , __UpperCamelCase =image.size else: __UpperCamelCase , __UpperCamelCase =image.shape[1], image.shape[2] if w < h: __UpperCamelCase =int(self.size['shortest_edge'] * h / w ) __UpperCamelCase =self.size['shortest_edge'] elif w > h: __UpperCamelCase =self.size['shortest_edge'] __UpperCamelCase =int(self.size['shortest_edge'] * w / h ) else: __UpperCamelCase =self.size['shortest_edge'] __UpperCamelCase =self.size['shortest_edge'] else: __UpperCamelCase =[] for image in image_inputs: __UpperCamelCase , __UpperCamelCase =self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __UpperCamelCase =max(A_ , key=lambda A_ : item[0] )[0] __UpperCamelCase =max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCAmelCase__ ( A_ , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : int = DeformableDetrImageProcessor if is_vision_available() else None def _a ( self ) -> str: __UpperCamelCase =DeformableDetrImageProcessingTester(self ) @property def _a ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def _a ( self ) -> int: __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) def _a ( self ) -> str: __UpperCamelCase =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) __UpperCamelCase =self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def _a ( self ) -> Optional[int]: pass def _a ( self ) -> Tuple: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random PIL images __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ , batched=A_ ) __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _a ( self ) -> Dict: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _a ( self ) -> List[Any]: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values __UpperCamelCase , __UpperCamelCase =self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _a ( self ) -> Optional[Any]: # prepare image and target __UpperCamelCase =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __UpperCamelCase =json.loads(f.read() ) __UpperCamelCase ={'image_id': 39769, 'annotations': target} # encode them __UpperCamelCase =DeformableDetrImageProcessor() __UpperCamelCase =image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values __UpperCamelCase =torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) __UpperCamelCase =torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area __UpperCamelCase =torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes __UpperCamelCase =torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) __UpperCamelCase =torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id __UpperCamelCase =torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd __UpperCamelCase =torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels __UpperCamelCase =torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size __UpperCamelCase =torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size __UpperCamelCase =torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def _a ( self ) -> List[Any]: # prepare image, target and masks_path __UpperCamelCase =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __UpperCamelCase =json.loads(f.read() ) __UpperCamelCase ={'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} __UpperCamelCase =pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __UpperCamelCase =DeformableDetrImageProcessor(format='coco_panoptic' ) __UpperCamelCase =image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values __UpperCamelCase =torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) __UpperCamelCase =torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area __UpperCamelCase =torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes __UpperCamelCase =torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) __UpperCamelCase =torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id __UpperCamelCase =torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd __UpperCamelCase =torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels __UpperCamelCase =torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks __UpperCamelCase =822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size __UpperCamelCase =torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size __UpperCamelCase =torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
62
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class UpperCAmelCase__ ( A_ , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : Tuple = TransfoXLTokenizer UpperCAmelCase__ : str = False UpperCAmelCase__ : Tuple = False def _a ( self ) -> Union[str, Any]: super().setUp() __UpperCamelCase =[ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _a ( self , **A_ ) -> Optional[int]: __UpperCamelCase =True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def _a ( self , A_ ) -> Tuple: __UpperCamelCase ='<unk> UNwanted , running' __UpperCamelCase ='<unk> unwanted, running' return input_text, output_text def _a ( self ) -> str: __UpperCamelCase =TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) __UpperCamelCase =tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def _a ( self ) -> Any: __UpperCamelCase =TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def _a ( self ) -> Optional[int]: __UpperCamelCase =TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def _a ( self ) -> int: __UpperCamelCase =TransfoXLTokenizer(lower_case=A_ ) __UpperCamelCase ='Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' __UpperCamelCase =[ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def _a ( self ) -> Optional[int]: __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
62
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase (lowercase_: Optional[Any] ) -> int: A__ : Tuple = [2, 2, 6, 2] if """tiny""" in model_name else [2, 2, 18, 2] A__ : Optional[Any] = True if """large""" in model_name or """huge""" in model_name else False A__ : Optional[int] = True if """large""" in model_name or """huge""" in model_name else False A__ : Optional[Any] = True if """large""" in model_name or """huge""" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: A__ : List[str] = [3, 3, 3, 3] A__ : List[str] = [5, 5, 5, 5] elif "fl4" in model_name: A__ : List[str] = [4, 4, 4, 4] A__ : Optional[Any] = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: A__ : Optional[int] = [3, 3, 3, 3] if "lrf" in model_name: A__ : Optional[Any] = [3, 3, 3, 3] else: A__ : Tuple = [2, 2, 2, 2] if "tiny" in model_name: A__ : str = 96 elif "small" in model_name: A__ : Dict = 96 elif "base" in model_name: A__ : Any = 128 elif "large" in model_name: A__ : List[str] = 192 elif "xlarge" in model_name: A__ : Tuple = 256 elif "huge" in model_name: A__ : List[Any] = 352 # set label information A__ : Optional[int] = """huggingface/label-files""" if "large" in model_name or "huge" in model_name: A__ : Optional[int] = """imagenet-22k-id2label.json""" else: A__ : Optional[int] = """imagenet-1k-id2label.json""" A__ : str = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="""dataset""" ) , """r""" ) ) A__ : Optional[Any] = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} A__ : int = {v: k for k, v in idalabel.items()} A__ : str = FocalNetConfig( embed_dim=SCREAMING_SNAKE_CASE_ , depths=SCREAMING_SNAKE_CASE_ , focal_levels=SCREAMING_SNAKE_CASE_ , focal_windows=SCREAMING_SNAKE_CASE_ , use_conv_embed=SCREAMING_SNAKE_CASE_ , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , use_post_layernorm=SCREAMING_SNAKE_CASE_ , use_layerscale=SCREAMING_SNAKE_CASE_ , ) return config def UpperCamelCase (lowercase_: Union[str, Any] ) -> Optional[int]: if "patch_embed.proj" in name: A__ : Union[str, Any] = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: A__ : str = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: A__ : Any = """encoder.""" + name if "encoder.layers" in name: A__ : int = name.replace("""encoder.layers""" , """encoder.stages""" ) if "downsample.proj" in name: A__ : int = name.replace("""downsample.proj""" , """downsample.projection""" ) if "blocks" in name: A__ : str = name.replace("""blocks""" , """layers""" ) if "modulation.f.weight" in name or "modulation.f.bias" in name: A__ : int = name.replace("""modulation.f""" , """modulation.projection_in""" ) if "modulation.h.weight" in name or "modulation.h.bias" in name: A__ : str = name.replace("""modulation.h""" , """modulation.projection_context""" ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: A__ : List[str] = name.replace("""modulation.proj""" , """modulation.projection_out""" ) if name == "norm.weight": A__ : List[str] = """layernorm.weight""" if name == "norm.bias": A__ : int = """layernorm.bias""" if "head" in name: A__ : int = name.replace("""head""" , """classifier""" ) else: A__ : int = """focalnet.""" + name return name def UpperCamelCase (lowercase_: Optional[int] , lowercase_: List[Any] , lowercase_: Optional[Any]=False ) -> Dict: # fmt: off A__ : int = { """focalnet-tiny""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth""", """focalnet-tiny-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth""", """focalnet-small""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth""", """focalnet-small-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth""", """focalnet-base""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth""", """focalnet-base-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth""", """focalnet-large-lrf-fl3""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth""", """focalnet-large-lrf-fl4""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth""", """focalnet-xlarge-lrf-fl3""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth""", """focalnet-xlarge-lrf-fl4""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth""", } # fmt: on A__ : Dict = model_name_to_url[model_name] print("""Checkpoint URL: """ , SCREAMING_SNAKE_CASE_ ) A__ : int = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE_ , map_location="""cpu""" )["""model"""] # rename keys for key in state_dict.copy().keys(): A__ : List[str] = state_dict.pop(SCREAMING_SNAKE_CASE_ ) A__ : Optional[int] = val A__ : List[str] = get_focalnet_config(SCREAMING_SNAKE_CASE_ ) A__ : Any = FocalNetForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() # load state dict model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # verify conversion A__ : Optional[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : Any = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"""shortest_edge""": 256} , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size=224 , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=SCREAMING_SNAKE_CASE_ , image_std=SCREAMING_SNAKE_CASE_ , ) A__ : str = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) A__ : Dict = processor(images=SCREAMING_SNAKE_CASE_ , return_tensors="""pt""" ) A__ : Union[str, Any] = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) A__ : Union[str, Any] = image_transforms(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) A__ : Tuple = model(**SCREAMING_SNAKE_CASE_ ) A__ : List[str] = outputs.logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) print("""First values of logits:""" , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": A__ : Optional[Any] = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": A__ : List[str] = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": A__ : int = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": A__ : List[str] = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": A__ : str = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": A__ : int = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f"""Saving model and processor of {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f"""Pushing model and processor of {model_name} to the hub...""" ) model.push_to_hub(f"""{model_name}""" ) processor.push_to_hub(f"""{model_name}""" ) if __name__ == "__main__": A_ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) A_ : List[str] = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
364
import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class _a : '''simple docstring''' def __init__( self , A__ , A__=3 , A__=7 , A__=True , A__=True , A__=False , A__=True , A__=99 , A__=32 , A__=5 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=512 , A__=16 , A__=2 , A__=0.0_2 , A__=3 , A__=4 , A__=None , ): A__ : List[Any] = parent A__ : List[str] = batch_size A__ : Optional[int] = seq_length A__ : Optional[int] = is_training A__ : Any = use_input_mask A__ : Tuple = use_token_type_ids A__ : str = use_labels A__ : Tuple = vocab_size A__ : Any = hidden_size A__ : List[str] = num_hidden_layers A__ : Optional[int] = num_attention_heads A__ : Optional[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Union[str, Any] = attention_probs_dropout_prob A__ : List[str] = max_position_embeddings A__ : Union[str, Any] = type_vocab_size A__ : str = type_sequence_label_size A__ : Tuple = initializer_range A__ : Tuple = num_labels A__ : Dict = num_choices A__ : List[str] = scope def __A ( self ): A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Any = None if self.use_input_mask: A__ : int = random_attention_mask([self.batch_size, self.seq_length] ) A__ : str = None A__ : Union[str, Any] = None A__ : List[str] = None A__ : Optional[Any] = None if self.use_labels: A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) A__ : Dict = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __A ( self ): return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A__ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=A__ , ) def __A ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ ): A__ : List[str] = FalconModel(config=A__ ) model.to(A__ ) model.eval() A__ : int = model(A__ , attention_mask=A__ ) A__ : Union[str, Any] = model(A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): A__ : Union[str, Any] = True A__ : Union[str, Any] = FalconModel(A__ ) model.to(A__ ) model.eval() A__ : Tuple = model( A__ , attention_mask=A__ , encoder_hidden_states=A__ , encoder_attention_mask=A__ , ) A__ : Union[str, Any] = model( A__ , attention_mask=A__ , encoder_hidden_states=A__ , ) A__ : List[str] = model(A__ , attention_mask=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): A__ : Any = FalconForCausalLM(config=A__ ) model.to(A__ ) model.eval() A__ : Tuple = model(A__ , attention_mask=A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __A ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): A__ : Optional[Any] = True A__ : Union[str, Any] = True A__ : int = FalconForCausalLM(config=A__ ) model.to(A__ ) model.eval() # first forward pass A__ : List[Any] = model( A__ , attention_mask=A__ , encoder_hidden_states=A__ , encoder_attention_mask=A__ , use_cache=A__ , ) A__ : Tuple = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Tuple = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : List[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[str] = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Optional[int] = model( A__ , attention_mask=A__ , encoder_hidden_states=A__ , encoder_attention_mask=A__ , output_hidden_states=A__ , )["""hidden_states"""][0] A__ : Any = model( A__ , attention_mask=A__ , encoder_hidden_states=A__ , encoder_attention_mask=A__ , past_key_values=A__ , output_hidden_states=A__ , )["""hidden_states"""][0] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A__ , A__ , atol=1e-3 ) ) def __A ( self ): A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Tuple = config_and_inputs A__ : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _a (__magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: List[Any] = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) UpperCAmelCase__: Tuple = (FalconForCausalLM,) if is_torch_available() else () UpperCAmelCase__: Optional[int] = ( { '''feature-extraction''': FalconModel, '''text-classification''': FalconForSequenceClassification, '''text-generation''': FalconForCausalLM, '''question-answering''': FalconForQuestionAnswering, '''token-classification''': FalconForTokenClassification, '''zero-shot''': FalconForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__: str = False UpperCAmelCase__: int = False def __A ( self ): A__ : List[Any] = FalconModelTester(self ) A__ : Union[str, Any] = ConfigTester(self , config_class=A__ , hidden_size=37 ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): A__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def __A ( self ): A__ , *A__ : List[Any] = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: A__ : Tuple = alibi self.model_tester.create_and_check_model(A__ , *A__ ) def __A ( self ): A__ , A__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : int = input_dict["""input_ids"""] A__ : int = input_ids.ne(1 ).to(A__ ) A__ : Union[str, Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Optional[int] = FalconForSequenceClassification(A__ ) model.to(A__ ) model.eval() A__ : int = model(A__ , attention_mask=A__ , labels=A__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __A ( self ): A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Dict = 3 A__ : Tuple = """single_label_classification""" A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(A__ ) A__ : Dict = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Any = FalconForSequenceClassification(A__ ) model.to(A__ ) model.eval() A__ : Any = model(A__ , attention_mask=A__ , labels=A__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __A ( self ): A__ , A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[str] = input_dict["""input_ids"""] A__ : List[str] = FalconForCausalLM(A__ ) model.to(A__ ) model.eval() A__ : Any = model(A__ , use_cache=A__ ) A__ : Any = input_ids.shape[0] A__ : Union[str, Any] = model._convert_to_rw_cache(result.past_key_values ) A__ : int = model._convert_cache_to_standard_format(A__ , A__ ) for layer in range(len(A__ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def __A ( self ): A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[Any] = 3 A__ : List[Any] = """multi_label_classification""" A__ : Tuple = input_dict["""input_ids"""] A__ : List[Any] = input_ids.ne(1 ).to(A__ ) A__ : Optional[Any] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Optional[int] = FalconForSequenceClassification(A__ ) model.to(A__ ) model.eval() A__ : List[Any] = model(A__ , attention_mask=A__ , labels=A__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __A ( self ): # Falcon can have different numbers of KV-heads than the number of query heads, so we need # to override this test to use the right head counts. for model_class in self.all_generative_model_classes: A__ , A__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(A__ , """use_cache""" ): return A__ : Optional[Any] = model_class(A__ ).to(A__ ) if "use_cache" not in inputs: A__ : Optional[int] = True A__ : List[Any] = model(**A__ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return A__ : str = ( getattr(A__ , """decoder_layers""" , A__ ) or getattr(A__ , """num_decoder_layers""" , A__ ) or config.num_hidden_layers ) A__ : Dict = getattr(A__ , """num_kv_heads""" , config.num_attention_heads ) A__ : List[str] = getattr(A__ , """d_model""" , config.hidden_size ) A__ : Union[str, Any] = embed_dim // num_attention_heads A__ : str = outputs["""past_key_values"""] self.assertEqual(len(A__ ) , A__ ) A__ , A__ : int = inputs["""input_ids"""].shape for i in range(A__ ): if config.new_decoder_architecture: A__ : Any = config.num_attention_heads elif config.multi_query: A__ : List[Any] = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class _a (unittest.TestCase ): '''simple docstring''' @slow def __A ( self ): A__ : Dict = AutoTokenizer.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) A__ : List[Any] = FalconForCausalLM.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) model.eval() model.to(A__ ) A__ : Optional[int] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(A__ ) A__ : Optional[Any] = ( """My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday.""" ) A__ : Any = model.generate(**A__ , do_sample=A__ , max_new_tokens=19 ) A__ : Optional[int] = tokenizer.batch_decode(A__ )[0] self.assertEqual(A__ , A__ ) @slow def __A ( self ): # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: A__ : Dict = AutoTokenizer.from_pretrained(A__ ) A__ : List[str] = FalconForCausalLM.from_pretrained(A__ ) model.eval() model.to(A__ ) A__ : Union[str, Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(A__ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**A__ , do_sample=A__ , max_new_tokens=4 ) model.generate(**A__ , do_sample=A__ , max_new_tokens=4 ) model.generate(**A__ , num_beams=2 , max_new_tokens=4 ) @slow def __A ( self ): # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: A__ : Dict = AutoTokenizer.from_pretrained(A__ ) A__ : Any = FalconForCausalLM.from_pretrained(A__ ) model.eval() model.to(device=A__ ) A__ : List[str] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(A__ ) # Test results are the same with and without cache A__ : Tuple = model.generate(**A__ , do_sample=A__ , max_new_tokens=20 , use_cache=A__ ) A__ : Optional[Any] = model.generate(**A__ , do_sample=A__ , max_new_tokens=20 , use_cache=A__ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
141
0
import math def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: '''simple docstring''' if ( not isinstance(UpperCamelCase__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * power_factor def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' if ( not isinstance(UpperCamelCase__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
273
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = len(lowerCAmelCase ) for i in range(length - 1 ): _lowerCAmelCase = i for k in range(i + 1 , lowerCAmelCase ): if collection[k] < collection[least]: _lowerCAmelCase = k if least != i: _lowerCAmelCase , _lowerCAmelCase = (collection[i], collection[least]) return collection if __name__ == "__main__": A__ : str =input('''Enter numbers separated by a comma:\n''').strip() A__ : Optional[int] =[int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
70
0
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): __a = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] __a = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } __a = f'{src_lang}-{tgt_lang}' __a = f'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) __a = os.path.join(_UpperCAmelCase , '''README.md''' ) print(f'Generating {path}' ) with open(_UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(_UpperCAmelCase ) # make sure we are under the root of the project __snake_case :List[Any] = Path(__file__).resolve().parent.parent.parent __snake_case :int = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __snake_case :Optional[Any] = model_name.split('''-''') __snake_case :Dict = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
363
def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): return base * power(_UpperCAmelCase , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('''Raise base to the power of exponent using recursion...''') __snake_case :List[Any] = int(input('''Enter the base: ''').strip()) __snake_case :Dict = int(input('''Enter the exponent: ''').strip()) __snake_case :int = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents __snake_case :Optional[Any] = 1 / result print(f'{base} to the power of {exponent} is {result}')
131
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class A__ ( unittest.TestCase): @slow def __lowerCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Tuple = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = TFAutoModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = AutoModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCAmelCase : int = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = TFAutoModelForPreTraining.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : str = AutoModelForPreTraining.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Dict = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = TFAutoModelForCausalLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : List[str] = TFAutoModelForCausalLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Any = AutoModelForCausalLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : List[Any] = AutoModelForCausalLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Dict = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = TFAutoModelForMaskedLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : int = TFAutoModelForMaskedLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = AutoModelForMaskedLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : str = AutoModelForMaskedLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[int] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : Dict = AutoModelForSeqaSeqLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Dict = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = TFAutoModelForSequenceClassification.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = TFAutoModelForQuestionAnswering.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = AutoModelForQuestionAnswering.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): __lowerCAmelCase : List[Any] = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE ) , 1_44_10 ) __lowerCAmelCase : Tuple = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE ) , 1_44_10 ) def __lowerCamelCase ( self ): __lowerCAmelCase : int = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE ) , 1_44_10 ) __lowerCAmelCase : Tuple = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE ) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE ) , 1_44_10 )
86
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): return round(float(moles / volume ) * nfactor ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): return round(float((moles * 0.0821 * temperature) / (volume) ) ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
86
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) _lowerCAmelCase = { "configuration_trocr": ["TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP", "TrOCRConfig"], "processing_trocr": ["TrOCRProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "TROCR_PRETRAINED_MODEL_ARCHIVE_LIST", "TrOCRForCausalLM", "TrOCRPreTrainedModel", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
367
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCAmelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ["GLPNFeatureExtractor"] _lowerCAmelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
98
0
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["pixel_values"] def __init__( self , UpperCAmelCase = True , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = 8 , **UpperCAmelCase , ): """simple docstring""" super().__init__(**UpperCAmelCase ) _UpperCAmelCase = do_rescale _UpperCAmelCase = rescale_factor _UpperCAmelCase = do_pad _UpperCAmelCase = pad_size def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = get_image_size(UpperCAmelCase ) _UpperCAmelCase = (old_height // size + 1) * size - old_height _UpperCAmelCase = (old_width // size + 1) * size - old_width return pad(UpperCAmelCase , ((0, pad_height), (0, pad_width)) , mode='symmetric' , data_format=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ): """simple docstring""" _UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCAmelCase = do_pad if do_pad is not None else self.do_pad _UpperCAmelCase = pad_size if pad_size is not None else self.pad_size _UpperCAmelCase = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) # All transformations expect numpy arrays. _UpperCAmelCase = [to_numpy_array(UpperCAmelCase ) for image in images] if do_rescale: _UpperCAmelCase = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_pad: _UpperCAmelCase = [self.pad(UpperCAmelCase , size=UpperCAmelCase ) for image in images] _UpperCAmelCase = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _UpperCAmelCase = {'pixel_values': images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
39
'''simple docstring''' import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values lowercase_ = argparse.ArgumentParser() parser.add_argument("""--user""", type=str, default="""ubuntu""") parser.add_argument("""--host""", type=str, default="""localhost""") parser.add_argument("""--key_path""", type=str, default=None) parser.add_argument("""--instance""", type=str, default="""V100:1""") parser.add_argument("""--provider""", type=str, default="""cheapest""") parser.add_argument("""--use_spot""", type=bool, default=False) parser.add_argument("""--example""", type=str, default="""pytorch/text-generation/run_generation.py""") lowercase_ , lowercase_ = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("""Cannot specify both BYO and on-demand cluster args""") lowercase_ = rh.cluster( name="""rh-cluster""", ips=[args.host], ssh_creds={"""ssh_user""": args.user, """ssh_private_key""": args.key_path} ) else: lowercase_ = rh.cluster( name="""rh-cluster""", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) lowercase_ = args.example.rsplit("""/""", 1)[0] # Set up remote environment cluster.install_packages(["""pip:./"""]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f"""pip install -r transformers/examples/{example_dir}/requirements.txt"""]) cluster.run(["""pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"""]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f"""python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}"""]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
58
0
'''simple docstring''' from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass lowerCAmelCase : Any =(3, 9, -11, 0, 7, 5, 1, -1) lowerCAmelCase : List[str] =(4, 6, 2, 0, 8, 10, 3, -2) @dataclass class a_ : __A = 42 __A = 42 class a_ : def __init__( self : Tuple , lowercase : Optional[Any] ): """simple docstring""" lowercase_ :Node | None = None for i in sorted(SCREAMING_SNAKE_CASE_ , reverse=SCREAMING_SNAKE_CASE_ ): lowercase_ :Any = Node(SCREAMING_SNAKE_CASE_ , self.head ) def __iter__( self : Tuple ): """simple docstring""" lowercase_ :List[Any] = self.head while node: yield node.data lowercase_ :Optional[Any] = node.next_node def __len__( self : List[Any] ): """simple docstring""" return sum(1 for _ in self ) def __str__( self : Union[str, Any] ): """simple docstring""" return " -> ".join([str(SCREAMING_SNAKE_CASE_ ) for node in self] ) def UpperCAmelCase_ ( __lowerCamelCase : SortedLinkedList ,__lowerCamelCase : SortedLinkedList ): return SortedLinkedList(list(lowerCAmelCase__ ) + list(lowerCAmelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase : str =SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
366
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : list ): if len(__lowerCamelCase ) <= 1: return lst lowercase_ :Optional[Any] = 1 while i < len(__lowerCamelCase ): if lst[i - 1] <= lst[i]: i += 1 else: lowercase_ , lowercase_ :int = lst[i], lst[i - 1] i -= 1 if i == 0: lowercase_ :Dict = 1 return lst if __name__ == "__main__": lowerCAmelCase : Any =input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase : List[str] =[int(item) for item in user_input.split(''',''')] print(gnome_sort(unsorted))
147
0
'''simple docstring''' import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase =logging.get_logger(__name__) __UpperCAmelCase ={ "vocab_file": "vocab.json", "tokenizer_config_file": "tokenizer_config.json", "merges_file": "merges.txt", } __UpperCAmelCase ={ "vocab_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json" ), }, "tokenizer_config_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json" ), }, "merges_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt" ), }, } __UpperCAmelCase ="</w>" __UpperCAmelCase ="@@ " def __lowerCAmelCase ( UpperCamelCase__ ) -> str: __lowerCamelCase = set() __lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCamelCase = char return pairs # Speech2Text2 has no max input length __UpperCAmelCase ={"facebook/s2t-wav2vec2-large-en-de": 1_0_2_4} class a__ ( UpperCAmelCase__ ): lowerCamelCase : Optional[int] =VOCAB_FILES_NAMES lowerCamelCase : Optional[int] =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Any =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =["input_ids", "attention_mask"] def __init__( self : int , a : List[str] , a : int="<s>" , a : Any="<pad>" , a : Union[str, Any]="</s>" , a : Optional[Any]="<unk>" , a : List[Any]=False , a : Tuple=None , **a : str , ): """simple docstring""" super().__init__( unk_token=a , bos_token=a , eos_token=a , pad_token=a , do_lower_case=a , **a , ) __lowerCamelCase = do_lower_case with open(a , encoding='''utf-8''' ) as vocab_handle: __lowerCamelCase = json.load(a ) __lowerCamelCase = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) __lowerCamelCase = None __lowerCamelCase = None else: with open(a , encoding='''utf-8''' ) as merges_handle: __lowerCamelCase = merges_handle.read().split('''\n''' )[:-1] __lowerCamelCase = [tuple(merge.split()[:2] ) for merge in merges] __lowerCamelCase = dict(zip(a , range(len(a ) ) ) ) __lowerCamelCase = {} @property def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return len(self.decoder ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE__ ( self : Any , a : List[str] ): """simple docstring""" __lowerCamelCase = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] __lowerCamelCase = get_pairs(a ) if not pairs: return token while True: __lowerCamelCase = min(a , key=lambda a : self.bpe_ranks.get(a , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break __lowerCamelCase , __lowerCamelCase = bigram __lowerCamelCase = [] __lowerCamelCase = 0 while i < len(a ): try: __lowerCamelCase = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __lowerCamelCase = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowerCamelCase = tuple(a ) __lowerCamelCase = new_word if len(a ) == 1: break else: __lowerCamelCase = get_pairs(a ) __lowerCamelCase = ''' '''.join(a ) if word == "\n " + BPE_TOKEN_MERGES: __lowerCamelCase = '''\n''' + BPE_TOKEN_MERGES if word.endswith(a ): __lowerCamelCase = word.replace(a , '''''' ) __lowerCamelCase = word.replace(''' ''' , a ) __lowerCamelCase = word return word def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Any ): """simple docstring""" if self.bpe_ranks is None: raise ValueError( '''This tokenizer was instantiated without a `merges.txt` file, so''' ''' that it can only be used for decoding, not for encoding.''' '''Make sure to provide `merges.txt` file at instantiation to enable ''' '''encoding.''' ) if self.do_lower_case: __lowerCamelCase = text.lower() __lowerCamelCase = text.split() __lowerCamelCase = [] for token in text: if token: split_tokens.extend(list(self.bpe(a ).split(''' ''' ) ) ) return split_tokens def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : str ): """simple docstring""" return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : int ): """simple docstring""" __lowerCamelCase = self.decoder.get(a , self.unk_token ) return result def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , a : List[str] ): """simple docstring""" __lowerCamelCase = ''' '''.join(a ) # make sure @@ tokens are concatenated __lowerCamelCase = ''''''.join(string.split(a ) ) return string def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : str , a : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join( a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(a , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '''\n''' ) __lowerCamelCase = 0 if self.bpe_ranks is None: return (vocab_file,) with open(a , '''w''' , encoding='''utf-8''' ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) __lowerCamelCase = token_index writer.write(''' '''.join(a ) + '''\n''' ) index += 1 return (vocab_file, merges_file)
67
'''simple docstring''' import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging __UpperCAmelCase =logging.get_logger(__name__) def __lowerCAmelCase ( UpperCamelCase__=None , UpperCamelCase__=None ) -> int: return field(default_factory=lambda: default , metadata=UpperCamelCase__ ) @dataclass class a__ : lowerCamelCase : List[str] =list_field( default=[] , metadata={ "help": ( "Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version" " of all available models" ) } , ) lowerCamelCase : List[int] =list_field( default=[8] , metadata={"help": "List of batch sizes for which memory and time performance will be evaluated"} ) lowerCamelCase : List[int] =list_field( default=[8, 3_2, 1_2_8, 5_1_2] , metadata={"help": "List of sequence lengths for which memory and time performance will be evaluated"} , ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={"help": "Whether to benchmark inference of model. Inference can be disabled via --no-inference."} , ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={"help": "Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."} , ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={"help": "Whether to run on available tpu devices. TPU can be disabled via --no-tpu."} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Use FP16 to accelerate inference."} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Benchmark training of model"} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Verbose memory tracing"} ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={"help": "Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."} , ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={ "help": "Whether to perform memory measurements. Memory measurements can be disabled via --no-memory" } , ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Trace memory line by line"} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Save result to a CSV file"} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Save all print statements in a log file"} ) lowerCamelCase : bool =field(default=UpperCAmelCase__ , metadata={"help": "Whether to print environment information"} ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={ "help": ( "Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use" " multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled" " for debugging / testing and on TPU." ) } , ) lowerCamelCase : str =field( default=F'''inference_time_{round(time() )}.csv''' , metadata={"help": "CSV filename used if saving time results to csv."} , ) lowerCamelCase : str =field( default=F'''inference_memory_{round(time() )}.csv''' , metadata={"help": "CSV filename used if saving memory results to csv."} , ) lowerCamelCase : str =field( default=F'''train_time_{round(time() )}.csv''' , metadata={"help": "CSV filename used if saving time results to csv for training."} , ) lowerCamelCase : str =field( default=F'''train_memory_{round(time() )}.csv''' , metadata={"help": "CSV filename used if saving memory results to csv for training."} , ) lowerCamelCase : str =field( default=F'''env_info_{round(time() )}.csv''' , metadata={"help": "CSV filename used if saving environment information."} , ) lowerCamelCase : str =field( default=F'''log_{round(time() )}.csv''' , metadata={"help": "Log filename used if print statements are saved in log."} , ) lowerCamelCase : int =field(default=3 , metadata={"help": "Times an experiment will be run."} ) lowerCamelCase : bool =field( default=UpperCAmelCase__ , metadata={ "help": ( "Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain" " model weights." ) } , ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" warnings.warn( f"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" ''' are deprecated in general and it is advised to use external Benchmarking libraries ''' ''' to benchmark Transformer models.''' , a , ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" if len(self.models ) <= 0: raise ValueError( '''Please make sure you provide at least one model name / model identifier, *e.g.* `--models''' ''' bert-base-cased` or `args.models = [\'bert-base-cased\'].''' ) return self.models @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" if not self.multi_process: return False elif self.is_tpu: logger.info('''Multiprocessing is currently not possible on TPU.''' ) return False else: return True
67
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : Any = logging.get_logger(__name__) lowercase : Optional[int] = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json''' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A : int = 'trocr' A : Optional[int] = ['past_key_values'] A : Optional[int] = { 'num_attention_heads': 'decoder_attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'decoder_layers', } def __init__( self , _SCREAMING_SNAKE_CASE=5_0265 , _SCREAMING_SNAKE_CASE=1024 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=4096 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=512 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE=2 , **_SCREAMING_SNAKE_CASE , ) -> Optional[int]: snake_case_ : Union[str, Any] = vocab_size snake_case_ : Dict = d_model snake_case_ : Dict = decoder_layers snake_case_ : Optional[Any] = decoder_attention_heads snake_case_ : List[str] = decoder_ffn_dim snake_case_ : Optional[Any] = activation_function snake_case_ : Any = max_position_embeddings snake_case_ : Optional[int] = dropout snake_case_ : int = attention_dropout snake_case_ : Tuple = activation_dropout snake_case_ : Dict = init_std snake_case_ : str = decoder_layerdrop snake_case_ : Optional[Any] = use_cache snake_case_ : int = scale_embedding snake_case_ : Optional[int] = use_learned_position_embeddings snake_case_ : Optional[int] = layernorm_embedding super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , )
36
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCAmelCase__ ( _a : str ): snake_case_ : str = FileLock(str(tmpdir / "foo.lock" ) ) snake_case_ : Optional[Any] = FileLock(str(tmpdir / "foo.lock" ) ) snake_case_ : Any = 0.01 with locka.acquire(): with pytest.raises(_a ): snake_case_ : Optional[int] = time.time() locka.acquire(_a ) assert time.time() - _start > timeout def lowerCAmelCase__ ( _a : Union[str, Any] ): snake_case_ : List[str] = "a" * 10_00 + ".lock" snake_case_ : Optional[int] = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(".lock" ) assert not locka._lock_file.endswith(_a ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 snake_case_ : int = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(_a ): locka.acquire(0 )
36
1
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __magic_name__ ( lowerCAmelCase_ , unittest.TestCase ): SCREAMING_SNAKE_CASE = VideoToVideoSDPipeline SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} SCREAMING_SNAKE_CASE = PipelineTesterMixin.required_optional_params - {'latents'} SCREAMING_SNAKE_CASE = False # No `output_type`. SCREAMING_SNAKE_CASE = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0 ) __a =UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=32 , attention_head_dim=4 , ) __a =DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) __a =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __a =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) __a =CLIPTextModel(__snake_case ) __a =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __a ={ 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def __magic_name__ ( self , __snake_case , __snake_case=0 ) -> Union[str, Any]: '''simple docstring''' # 3 frames __a =floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(__snake_case ) ).to(__snake_case ) if str(__snake_case ).startswith('mps' ): __a =torch.manual_seed(__snake_case ) else: __a =torch.Generator(device=__snake_case ).manual_seed(__snake_case ) __a ={ 'prompt': 'A painting of a squirrel eating a burger', 'video': video, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a ='cpu' # ensure determinism for the device-dependent torch.Generator __a =self.get_dummy_components() __a =VideoToVideoSDPipeline(**__snake_case ) __a =sd_pipe.to(__snake_case ) sd_pipe.set_progress_bar_config(disable=__snake_case ) __a =self.get_dummy_inputs(__snake_case ) __a ='np' __a =sd_pipe(**__snake_case ).frames __a =frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) __a =np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__snake_case , expected_max_diff=5e-3 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def __magic_name__ ( self ) -> Dict: '''simple docstring''' pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def __magic_name__ ( self ) -> Any: '''simple docstring''' pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def __magic_name__ ( self ) -> Tuple: '''simple docstring''' pass def __magic_name__ ( self ) -> Tuple: '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __magic_name__ ( unittest.TestCase ): def __magic_name__ ( self ) -> List[str]: '''simple docstring''' __a =VideoToVideoSDPipeline.from_pretrained('cerspense/zeroscope_v2_XL' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames __a =torch.Generator(device='cpu' ).manual_seed(0 ) __a =torch.randn((1, 10, 3, 1024, 576) , generator=__snake_case ) __a =video.to('cuda' ) __a ='Spiderman is surfing' __a =pipe(__snake_case , video=__snake_case , generator=__snake_case , num_inference_steps=3 , output_type='pt' ).frames __a =np.array([-1.045_8984, -1.127_9297, -0.966_3086, -0.9150_3906, -0.7509_7656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
218
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase_( _snake_case : int , _snake_case : int ): """simple docstring""" __a =old_name if "patch_embed" in old_name: __a , __a , __a =old_name.split('.' ) if layer == "0": __a =old_name.replace('0' , 'convolution1' ) elif layer == "1": __a =old_name.replace('1' , 'batchnorm_before' ) elif layer == "3": __a =old_name.replace('3' , 'convolution2' ) else: __a =old_name.replace('4' , 'batchnorm_after' ) if "network" in old_name and re.search(r'\d\.\d' , _snake_case ): __a =r'\b\d{2}\b' if bool(re.search(_snake_case , _snake_case ) ): __a =re.search(r'\d\.\d\d.' , _snake_case ).group() else: __a =re.search(r'\d\.\d.' , _snake_case ).group() if int(match[0] ) < 6: __a =old_name.replace(_snake_case , '' ) __a =trimmed_name.replace('network' , match[0] + '.meta4D_layers.blocks.' + match[2:-1] ) __a ='intermediate_stages.' + trimmed_name else: __a =old_name.replace(_snake_case , '' ) if int(match[2] ) < num_meta4D_last_stage: __a =trimmed_name.replace('network' , 'meta4D_layers.blocks.' + match[2] ) else: __a =str(int(match[2] ) - num_meta4D_last_stage ) __a =trimmed_name.replace('network' , 'meta3D_layers.blocks.' + layer_index ) if "norm1" in old_name: __a =trimmed_name.replace('norm1' , 'layernorm1' ) elif "norm2" in old_name: __a =trimmed_name.replace('norm2' , 'layernorm2' ) elif "fc1" in old_name: __a =trimmed_name.replace('fc1' , 'linear_in' ) elif "fc2" in old_name: __a =trimmed_name.replace('fc2' , 'linear_out' ) __a ='last_stage.' + trimmed_name elif "network" in old_name and re.search(r'.\d.' , _snake_case ): __a =old_name.replace('network' , 'intermediate_stages' ) if "fc" in new_name: __a =new_name.replace('fc' , 'convolution' ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): __a =new_name.replace('norm1' , 'batchnorm_before' ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): __a =new_name.replace('norm2' , 'batchnorm_after' ) if "proj" in new_name: __a =new_name.replace('proj' , 'projection' ) if "dist_head" in new_name: __a =new_name.replace('dist_head' , 'distillation_classifier' ) elif "head" in new_name: __a =new_name.replace('head' , 'classifier' ) elif "patch_embed" in new_name: __a ='efficientformer.' + new_name elif new_name == "norm.weight" or new_name == "norm.bias": __a =new_name.replace('norm' , 'layernorm' ) __a ='efficientformer.' + new_name else: __a ='efficientformer.encoder.' + new_name return new_name def UpperCamelCase_( _snake_case : List[str] , _snake_case : Dict ): """simple docstring""" for key in checkpoint.copy().keys(): __a =checkpoint.pop(_snake_case ) __a =val return checkpoint def UpperCamelCase_( ): """simple docstring""" __a ='http://images.cocodataset.org/val2017/000000039769.jpg' __a =Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image def UpperCamelCase_( _snake_case : Path , _snake_case : Path , _snake_case : Path , _snake_case : bool ): """simple docstring""" __a =torch.load(_snake_case , map_location='cpu' )['model'] __a =EfficientFormerConfig.from_json_file(_snake_case ) __a =EfficientFormerForImageClassificationWithTeacher(_snake_case ) __a ='_'.join(checkpoint_path.split('/' )[-1].split('.' )[0].split('_' )[:-1] ) __a =config.depths[-1] - config.num_metaad_blocks + 1 __a =convert_torch_checkpoint(_snake_case , _snake_case ) model.load_state_dict(_snake_case ) model.eval() __a ={ 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } # prepare image __a =prepare_img() __a =256 __a =224 __a =EfficientFormerImageProcessor( size={'shortest_edge': image_size} , crop_size={'height': crop_size, 'width': crop_size} , resample=pillow_resamplings['bicubic'] , ) __a =processor(images=_snake_case , return_tensors='pt' ).pixel_values # original processing pipeline __a =Compose( [ Resize(_snake_case , interpolation=pillow_resamplings['bicubic'] ), CenterCrop(_snake_case ), ToTensor(), Normalize(_snake_case , _snake_case ), ] ) __a =image_transforms(_snake_case ).unsqueeze(0 ) assert torch.allclose(_snake_case , _snake_case ) __a =model(_snake_case ) __a =outputs.logits __a =(1, 1000) if "l1" in model_name: __a =torch.Tensor( [-0.1_312, 0.4_353, -1.0_499, -0.5_124, 0.4_183, -0.6_793, -1.3_777, -0.0_893, -0.7_358, -2.4_328] ) assert torch.allclose(logits[0, :10] , _snake_case , atol=1e-3 ) assert logits.shape == expected_shape elif "l3" in model_name: __a =torch.Tensor( [-1.3_150, -1.5_456, -1.2_556, -0.8_496, -0.7_127, -0.7_897, -0.9_728, -0.3_052, 0.3_751, -0.3_127] ) assert torch.allclose(logits[0, :10] , _snake_case , atol=1e-3 ) assert logits.shape == expected_shape elif "l7" in model_name: __a =torch.Tensor( [-1.0_283, -1.4_131, -0.5_644, -1.3_115, -0.5_785, -1.2_049, -0.7_528, 0.1_992, -0.3_822, -0.0_878] ) assert logits.shape == expected_shape else: raise ValueError( F'Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7' ) # Save Checkpoints Path(_snake_case ).mkdir(exist_ok=_snake_case ) model.save_pretrained(_snake_case ) print(F'Checkpoint successfuly converted. Model saved at {pytorch_dump_path}' ) processor.save_pretrained(_snake_case ) print(F'Processor successfuly saved at {pytorch_dump_path}' ) if push_to_hub: print('Pushing model to the hub...' ) model.push_to_hub( repo_id=F'Bearnardd/{pytorch_dump_path}' , commit_message='Add model' , use_temp_dir=_snake_case , ) processor.push_to_hub( repo_id=F'Bearnardd/{pytorch_dump_path}' , commit_message='Add image processor' , use_temp_dir=_snake_case , ) if __name__ == "__main__": _lowerCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--pytorch_model_path", default=None, type=str, required=True, help="Path to EfficientFormer pytorch checkpoint.", ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The json file for EfficientFormer model config.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Push model and image processor to the hub") parser.add_argument( "--no-push_to_hub", dest="push_to_hub", action="store_false", help="Do not push model and image processor to the hub", ) parser.set_defaults(push_to_hub=True) _lowerCAmelCase : Union[str, Any] = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
218
1
"""simple docstring""" import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu _SCREAMING_SNAKE_CASE : List[str] = get_tests_dir() + """/test_data/fsmt/fsmt_val_data.json""" with io.open(filename, """r""", encoding="""utf-8""") as f: _SCREAMING_SNAKE_CASE : Union[str, Any] = json.load(f) @require_torch class __a ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : Dict , lowercase_ : Optional[int] ): return FSMTTokenizer.from_pretrained(lowercase_ ) def _lowerCAmelCase ( self : Dict , lowercase_ : List[Any] ): UpperCamelCase__ : Any =FSMTForConditionalGeneration.from_pretrained(lowercase_ ).to(lowercase_ ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['''en-ru''', 26.0], ['''ru-en''', 22.0], ['''en-de''', 22.0], ['''de-en''', 29.0], ] ) @slow def _lowerCAmelCase ( self : str , lowercase_ : List[str] , lowercase_ : Tuple ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality UpperCamelCase__ : Any =f'''facebook/wmt19-{pair}''' UpperCamelCase__ : int =self.get_tokenizer(lowercase_ ) UpperCamelCase__ : Any =self.get_model(lowercase_ ) UpperCamelCase__ : Union[str, Any] =bleu_data[pair]['''src'''] UpperCamelCase__ : Tuple =bleu_data[pair]['''tgt'''] UpperCamelCase__ : str =tokenizer(lowercase_ , return_tensors='''pt''' , truncation=lowercase_ , padding='''longest''' ).to(lowercase_ ) UpperCamelCase__ : List[Any] =model.generate( input_ids=batch.input_ids , num_beams=8 , ) UpperCamelCase__ : Union[str, Any] =tokenizer.batch_decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ ) UpperCamelCase__ : Optional[int] =calculate_bleu(lowercase_ , lowercase_ ) print(lowercase_ ) self.assertGreaterEqual(scores['''bleu'''] , lowercase_ )
352
"""simple docstring""" import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _SCREAMING_SNAKE_CASE : Tuple = logging.getLogger(__name__) @dataclass(frozen=snake_case__ ) class __a : """simple docstring""" SCREAMING_SNAKE_CASE_ = 42 SCREAMING_SNAKE_CASE_ = 42 SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None @dataclass(frozen=snake_case__ ) class __a : """simple docstring""" SCREAMING_SNAKE_CASE_ = 42 SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None if is_torch_available(): import torch from torch.utils.data import Dataset class __a ( snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = 42 def __init__( self : Optional[int] , lowercase_ : str , lowercase_ : PreTrainedTokenizer , lowercase_ : str , lowercase_ : Optional[int] = None , lowercase_ : Optional[int]=False , lowercase_ : bool = False , ): UpperCamelCase__ : Tuple =hans_processors[task]() UpperCamelCase__ : Union[str, Any] =os.path.join( lowercase_ , '''cached_{}_{}_{}_{}'''.format( '''dev''' if evaluate else '''train''' , tokenizer.__class__.__name__ , str(lowercase_ ) , lowercase_ , ) , ) UpperCamelCase__ : int =processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCamelCase__ , UpperCamelCase__ : Union[str, Any] =label_list[2], label_list[1] UpperCamelCase__ : List[Any] =label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. UpperCamelCase__ : Any =cached_features_file + '''.lock''' with FileLock(lowercase_ ): if os.path.exists(lowercase_ ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) UpperCamelCase__ : Optional[int] =torch.load(lowercase_ ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) UpperCamelCase__ : str =( processor.get_dev_examples(lowercase_ ) if evaluate else processor.get_train_examples(lowercase_ ) ) logger.info('''Training examples: %s''' , len(lowercase_ ) ) UpperCamelCase__ : Tuple =hans_convert_examples_to_features(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) logger.info('''Saving features into cached file %s''' , lowercase_ ) torch.save(self.features , lowercase_ ) def __len__( self : Union[str, Any] ): return len(self.features ) def __getitem__( self : Optional[int] , lowercase_ : Optional[Any] ): return self.features[i] def _lowerCAmelCase ( self : int ): return self.label_list if is_tf_available(): import tensorflow as tf class __a : """simple docstring""" SCREAMING_SNAKE_CASE_ = 42 def __init__( self : Any , lowercase_ : str , lowercase_ : PreTrainedTokenizer , lowercase_ : str , lowercase_ : Optional[int] = 128 , lowercase_ : Union[str, Any]=False , lowercase_ : bool = False , ): UpperCamelCase__ : Any =hans_processors[task]() UpperCamelCase__ : Tuple =processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCamelCase__ , UpperCamelCase__ : Tuple =label_list[2], label_list[1] UpperCamelCase__ : Union[str, Any] =label_list UpperCamelCase__ : Any =processor.get_dev_examples(lowercase_ ) if evaluate else processor.get_train_examples(lowercase_ ) UpperCamelCase__ : Union[str, Any] =hans_convert_examples_to_features(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='''convert examples to features''' ): if ex_index % 1_0000 == 0: logger.info('''Writing example %d of %d''' % (ex_index, len(lowercase_ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) UpperCamelCase__ : Optional[Any] =tf.data.Dataset.from_generator( lowercase_ , ( { '''example_id''': tf.intaa, '''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa, }, tf.intaa, ) , ( { '''example_id''': tf.TensorShape([] ), '''input_ids''': tf.TensorShape([None, None] ), '''attention_mask''': tf.TensorShape([None, None] ), '''token_type_ids''': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def _lowerCAmelCase ( self : Optional[Any] ): return self.dataset def __len__( self : str ): return len(self.features ) def __getitem__( self : List[str] , lowercase_ : Dict ): return self.features[i] def _lowerCAmelCase ( self : Dict ): return self.label_list class __a ( snake_case__ ): """simple docstring""" def _lowerCAmelCase ( self : List[Any] , lowercase_ : Union[str, Any] ): return self._create_examples(self._read_tsv(os.path.join(lowercase_ , '''heuristics_train_set.txt''' ) ) , '''train''' ) def _lowerCAmelCase ( self : Tuple , lowercase_ : Optional[int] ): return self._create_examples(self._read_tsv(os.path.join(lowercase_ , '''heuristics_evaluation_set.txt''' ) ) , '''dev''' ) def _lowerCAmelCase ( self : List[Any] ): return ["contradiction", "entailment", "neutral"] def _lowerCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : List[str] ): UpperCamelCase__ : Tuple =[] for i, line in enumerate(lowercase_ ): if i == 0: continue UpperCamelCase__ : str ='''%s-%s''' % (set_type, line[0]) UpperCamelCase__ : str =line[5] UpperCamelCase__ : Any =line[6] UpperCamelCase__ : Optional[int] =line[7][2:] if line[7].startswith('''ex''' ) else line[7] UpperCamelCase__ : str =line[0] examples.append(InputExample(guid=lowercase_ , text_a=lowercase_ , text_b=lowercase_ , label=lowercase_ , pairID=lowercase_ ) ) return examples def _lowerCAmelCase ( UpperCAmelCase : List[InputExample] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : PreTrainedTokenizer , ): '''simple docstring''' UpperCamelCase__ : List[str] ={label: i for i, label in enumerate(UpperCAmelCase )} UpperCamelCase__ : int =[] for ex_index, example in tqdm.tqdm(enumerate(UpperCAmelCase ) , desc='''convert examples to features''' ): if ex_index % 10_000 == 0: logger.info('''Writing example %d''' % (ex_index) ) UpperCamelCase__ : str =tokenizer( example.text_a , example.text_b , add_special_tokens=UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' , truncation=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , ) UpperCamelCase__ : str =label_map[example.label] if example.label in label_map else 0 UpperCamelCase__ : int =int(example.pairID ) features.append(InputFeatures(**UpperCAmelCase , label=UpperCAmelCase , pairID=UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('''*** Example ***''' ) logger.info(F'''guid: {example}''' ) logger.info(F'''features: {features[i]}''' ) return features _SCREAMING_SNAKE_CASE : List[str] = { """hans""": 3, } _SCREAMING_SNAKE_CASE : Tuple = { """hans""": HansProcessor, }
157
0
'''simple docstring''' from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def _lowerCamelCase ( lowercase : bool = True , *lowercase : Any , **lowercase : List[Any] ) -> Any: if not is_tqdm_available(): raise ImportError("Accelerate's `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`." ) _a = False if main_process_only: _a = PartialState().local_process_index == 0 return _tqdm(*lowercase , **lowercase , disable=lowercase )
63
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: _lowercase : int = None _lowercase : Union[str, Any] = logging.get_logger(__name__) _lowercase : Tuple = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} _lowercase : Dict = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", }, "tokenizer_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json", }, } _lowercase : int = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } _lowercase : Tuple = "▁" class __magic_name__ ( _UpperCAmelCase): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = AlbertTokenizer def __init__( self : Optional[Any] , lowercase_ : Union[str, Any]=None , lowercase_ : str=None , lowercase_ : Any=True , lowercase_ : Optional[int]=True , lowercase_ : List[str]=False , lowercase_ : Optional[int]="[CLS]" , lowercase_ : Any="[SEP]" , lowercase_ : int="<unk>" , lowercase_ : Any="[SEP]" , lowercase_ : int="<pad>" , lowercase_ : Tuple="[CLS]" , lowercase_ : Dict="[MASK]" , **lowercase_ : Optional[int] , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowercase_ : Tuple = ( AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ , normalized=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) else mask_token ) super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , remove_space=lowercase_ , keep_accents=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , **lowercase_ , ) lowercase_ : Optional[int] = do_lower_case lowercase_ : Any = remove_space lowercase_ : Dict = keep_accents lowercase_ : Optional[int] = vocab_file lowercase_ : Any = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ): lowercase_ : Tuple = [self.sep_token_id] lowercase_ : str = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE_ ( self : Any , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ): lowercase_ : Union[str, Any] = [self.sep_token_id] lowercase_ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , lowercase_ : str , lowercase_ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(lowercase_ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ : Optional[Any] = os.path.join( lowercase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase_ ): copyfile(self.vocab_file , lowercase_ ) return (out_vocab_file,)
239
0
"""simple docstring""" from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(_UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , **__a : int ) -> List[Any]: super().__init__(**__a ) if self.framework != "pt": raise ValueError(F'''The {self.__class__} is only available in PyTorch.''' ) # No specific FOR_XXX available yet def __call__( self : List[Any] , __a : Union[np.ndarray, bytes, str] , **__a : Tuple ) -> Any: return super().__call__(__a , **__a ) def __SCREAMING_SNAKE_CASE ( self : Optional[Any] , **__a : Union[str, Any] ) -> Dict: _UpperCamelCase : Tuple = {} if "candidate_labels" in kwargs: _UpperCamelCase : Optional[int] = kwargs["candidate_labels"] if "hypothesis_template" in kwargs: _UpperCamelCase : Optional[int] = kwargs["hypothesis_template"] return preprocess_params, {}, {} def __SCREAMING_SNAKE_CASE ( self : List[Any] , __a : Any , __a : str=None , __a : Dict="This is a sound of {}." ) -> Tuple: if isinstance(__a , __a ): if audio.startswith("http://" ) or audio.startswith("https://" ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png _UpperCamelCase : Optional[Any] = requests.get(__a ).content else: with open(__a , "rb" ) as f: _UpperCamelCase : Tuple = f.read() if isinstance(__a , __a ): _UpperCamelCase : Union[str, Any] = ffmpeg_read(__a , self.feature_extractor.sampling_rate ) if not isinstance(__a , np.ndarray ): raise ValueError("We expect a numpy ndarray as input" ) if len(audio.shape ) != 1: raise ValueError("We expect a single channel audio input for ZeroShotAudioClassificationPipeline" ) _UpperCamelCase : Union[str, Any] = self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors="pt" ) _UpperCamelCase : Tuple = candidate_labels _UpperCamelCase : int = [hypothesis_template.format(__a ) for x in candidate_labels] _UpperCamelCase : Optional[Any] = self.tokenizer(__a , return_tensors=self.framework , padding=__a ) _UpperCamelCase : Optional[int] = [text_inputs] return inputs def __SCREAMING_SNAKE_CASE ( self : List[str] , __a : int ) -> List[Any]: _UpperCamelCase : Optional[int] = model_inputs.pop("candidate_labels" ) _UpperCamelCase : str = model_inputs.pop("text_inputs" ) if isinstance(text_inputs[0] , __a ): _UpperCamelCase : List[Any] = text_inputs[0] else: # Batching case. _UpperCamelCase : int = text_inputs[0][0] _UpperCamelCase : Union[str, Any] = self.model(**__a , **__a ) _UpperCamelCase : Dict = { "candidate_labels": candidate_labels, "logits": outputs.logits_per_audio, } return model_outputs def __SCREAMING_SNAKE_CASE ( self : List[str] , __a : Optional[Any] ) -> Union[str, Any]: _UpperCamelCase : List[str] = model_outputs.pop("candidate_labels" ) _UpperCamelCase : int = model_outputs["logits"][0] if self.framework == "pt": _UpperCamelCase : Optional[Any] = logits.softmax(dim=0 ) _UpperCamelCase : Dict = probs.tolist() else: raise ValueError("`tf` framework not supported." ) _UpperCamelCase : Tuple = [ {"score": score, "label": candidate_label} for score, candidate_label in sorted(zip(__a , __a ) , key=lambda __a : -x[0] ) ] return result
359
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __SCREAMING_SNAKE_CASE : '''simple docstring''' @staticmethod def __SCREAMING_SNAKE_CASE ( *__a : int , **__a : int ) -> List[Any]: pass @is_pipeline_test @require_vision @require_timm @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :str = MODEL_FOR_OBJECT_DETECTION_MAPPING def __SCREAMING_SNAKE_CASE ( self : Any , __a : Union[str, Any] , __a : Optional[int] , __a : str ) -> Optional[Any]: _UpperCamelCase : List[Any] = ObjectDetectionPipeline(model=__a , image_processor=__a ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def __SCREAMING_SNAKE_CASE ( self : List[str] , __a : List[Any] , __a : Union[str, Any] ) -> int: _UpperCamelCase : Any = object_detector("./tests/fixtures/tests_samples/COCO/000000039769.png" , threshold=0.0 ) self.assertGreater(len(__a ) , 0 ) for detected_object in outputs: self.assertEqual( __a , { "score": ANY(__a ), "label": ANY(__a ), "box": {"xmin": ANY(__a ), "ymin": ANY(__a ), "xmax": ANY(__a ), "ymax": ANY(__a )}, } , ) import datasets _UpperCamelCase : str = datasets.load_dataset("hf-internal-testing/fixtures_image_utils" , "image" , split="test" ) _UpperCamelCase : List[Any] = [ Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ), "http://images.cocodataset.org/val2017/000000039769.jpg", # RGBA dataset[0]["file"], # LA dataset[1]["file"], # L dataset[2]["file"], ] _UpperCamelCase : List[Any] = object_detector(__a , threshold=0.0 ) self.assertEqual(len(__a ) , len(__a ) ) for outputs in batch_outputs: self.assertGreater(len(__a ) , 0 ) for detected_object in outputs: self.assertEqual( __a , { "score": ANY(__a ), "label": ANY(__a ), "box": {"xmin": ANY(__a ), "ymin": ANY(__a ), "xmax": ANY(__a ), "ymax": ANY(__a )}, } , ) @require_tf @unittest.skip("Object detection not implemented in TF" ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: pass @require_torch def __SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: _UpperCamelCase : List[str] = "hf-internal-testing/tiny-detr-mobilenetsv3" _UpperCamelCase : Optional[int] = AutoModelForObjectDetection.from_pretrained(__a ) _UpperCamelCase : str = AutoFeatureExtractor.from_pretrained(__a ) _UpperCamelCase : List[Any] = ObjectDetectionPipeline(model=__a , feature_extractor=__a ) _UpperCamelCase : int = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" , threshold=0.0 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ] , ) _UpperCamelCase : Any = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] , threshold=0.0 , ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ], [ {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.33_76, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ], ] , ) @require_torch @slow def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> Union[str, Any]: _UpperCamelCase : str = "facebook/detr-resnet-50" _UpperCamelCase : Union[str, Any] = AutoModelForObjectDetection.from_pretrained(__a ) _UpperCamelCase : str = AutoFeatureExtractor.from_pretrained(__a ) _UpperCamelCase : Union[str, Any] = ObjectDetectionPipeline(model=__a , feature_extractor=__a ) _UpperCamelCase : Tuple = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) _UpperCamelCase : List[str] = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], ] , ) @require_torch @slow def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> Union[str, Any]: _UpperCamelCase : Dict = "facebook/detr-resnet-50" _UpperCamelCase : Optional[Any] = pipeline("object-detection" , model=__a ) _UpperCamelCase : str = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) _UpperCamelCase : Tuple = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], [ {"score": 0.99_82, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.99_60, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.99_55, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], ] , ) @require_torch @slow def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: _UpperCamelCase : Tuple = 0.99_85 _UpperCamelCase : List[Any] = "facebook/detr-resnet-50" _UpperCamelCase : List[str] = pipeline("object-detection" , model=__a ) _UpperCamelCase : Any = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" , threshold=__a ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {"score": 0.99_88, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.99_87, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) @require_torch @require_pytesseract @slow def __SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: _UpperCamelCase : Optional[Any] = "Narsil/layoutlmv3-finetuned-funsd" _UpperCamelCase : int = 0.99_93 _UpperCamelCase : str = pipeline("object-detection" , model=__a , threshold=__a ) _UpperCamelCase : Union[str, Any] = object_detector( "https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png" ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {"score": 0.99_93, "label": "I-ANSWER", "box": {"xmin": 294, "ymin": 254, "xmax": 343, "ymax": 264}}, {"score": 0.99_93, "label": "I-ANSWER", "box": {"xmin": 294, "ymin": 254, "xmax": 343, "ymax": 264}}, ] , )
310
0
from __future__ import annotations def lowerCamelCase__ ( _A ): '''simple docstring''' create_state_space_tree(_A , [] , 0 , [0 for i in range(len(_A ) )] ) def lowerCamelCase__ ( _A , _A , _A , _A , ): '''simple docstring''' if index == len(_A ): print(_A ) return for i in range(len(_A ) ): if not index_used[i]: current_sequence.append(sequence[i] ) snake_case_ = True create_state_space_tree(_A , _A , index + 1 , _A ) current_sequence.pop() snake_case_ = False lowercase__ : list[int | str] = [3, 1, 2, 4] generate_all_permutations(sequence) lowercase__ : list[int | str] = ["A", "B", "C"] generate_all_permutations(sequence_a)
187
class UpperCAmelCase : '''simple docstring''' def __init__( self : Dict ): """simple docstring""" snake_case_ = {} # Mapping from char to TrieNode snake_case_ = False def snake_case__ ( self : Dict , __lowercase : list[str] ): """simple docstring""" for word in words: self.insert(__lowercase ) def snake_case__ ( self : List[str] , __lowercase : str ): """simple docstring""" snake_case_ = self for char in word: if char not in curr.nodes: snake_case_ = TrieNode() snake_case_ = curr.nodes[char] snake_case_ = True def snake_case__ ( self : List[Any] , __lowercase : str ): """simple docstring""" snake_case_ = self for char in word: if char not in curr.nodes: return False snake_case_ = curr.nodes[char] return curr.is_leaf def snake_case__ ( self : Optional[Any] , __lowercase : str ): """simple docstring""" def _delete(__lowercase : TrieNode , __lowercase : str , __lowercase : int ) -> bool: if index == len(__lowercase ): # If word does not exist if not curr.is_leaf: return False snake_case_ = False return len(curr.nodes ) == 0 snake_case_ = word[index] snake_case_ = curr.nodes.get(__lowercase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted snake_case_ = _delete(__lowercase , __lowercase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , __lowercase , 0 ) def lowerCamelCase__ ( _A , _A ): '''simple docstring''' if node.is_leaf: print(_A , end=" " ) for key, value in node.nodes.items(): print_words(_A , word + key ) def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = "banana bananas bandana band apple all beast".split() snake_case_ = TrieNode() root.insert_many(_A ) # print_words(root, "") assert all(root.find(_A ) for word in words ) assert root.find("banana" ) assert not root.find("bandanas" ) assert not root.find("apps" ) assert root.find("apple" ) assert root.find("all" ) root.delete("all" ) assert not root.find("all" ) root.delete("banana" ) assert not root.find("banana" ) assert root.find("bananas" ) return True def lowerCamelCase__ ( _A , _A ): '''simple docstring''' print(str(_A ) , "works!" if passes else "doesn't work :(" ) def lowerCamelCase__ ( ): '''simple docstring''' assert test_trie() def lowerCamelCase__ ( ): '''simple docstring''' print_results("Testing trie functionality" , test_trie() ) if __name__ == "__main__": main()
187
1
'''simple docstring''' from collections import defaultdict from math import ceil, sqrt def __magic_name__( lowerCamelCase = 1_0_0_0_0_0_0, lowerCamelCase = 1_0): __lowerCAmelCase = defaultdict(lowerCamelCase) for outer_width in range(3, (t_limit // 4) + 2): if outer_width * outer_width > t_limit: __lowerCAmelCase = max( ceil(sqrt(outer_width * outer_width - t_limit)), 1) else: __lowerCAmelCase = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(lowerCamelCase, outer_width - 1, 2): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 1_0) if __name__ == "__main__": print(f"""{solution() = }""")
9
'''simple docstring''' import argparse import datetime def __magic_name__( lowerCamelCase): __lowerCAmelCase = { '''0''': '''Sunday''', '''1''': '''Monday''', '''2''': '''Tuesday''', '''3''': '''Wednesday''', '''4''': '''Thursday''', '''5''': '''Friday''', '''6''': '''Saturday''', } __lowerCAmelCase = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(lowerCamelCase) < 1_1: raise ValueError('''Must be 10 characters long''') # Get month __lowerCAmelCase = int(date_input[0] + date_input[1]) # Validate if not 0 < m < 1_3: raise ValueError('''Month must be between 1 - 12''') __lowerCAmelCase = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError('''Date separator must be \'-\' or \'/\'''') # Get day __lowerCAmelCase = int(date_input[3] + date_input[4]) # Validate if not 0 < d < 3_2: raise ValueError('''Date must be between 1 - 31''') # Get second separator __lowerCAmelCase = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError('''Date separator must be \'-\' or \'/\'''') # Get year __lowerCAmelCase = int(date_input[6] + date_input[7] + date_input[8] + date_input[9]) # Arbitrary year range if not 4_5 < y < 8_5_0_0: raise ValueError( '''Year out of range. There has to be some sort of limit...right?''') # Get datetime obj for validation __lowerCAmelCase = datetime.date(int(lowerCamelCase), int(lowerCamelCase), int(lowerCamelCase)) # Start math if m <= 2: __lowerCAmelCase = y - 1 __lowerCAmelCase = m + 1_2 # maths var __lowerCAmelCase = int(str(lowerCamelCase)[:2]) __lowerCAmelCase = int(str(lowerCamelCase)[2:]) __lowerCAmelCase = int(2.6 * m - 5.39) __lowerCAmelCase = int(c / 4) __lowerCAmelCase = int(k / 4) __lowerCAmelCase = int(d + k) __lowerCAmelCase = int(t + u + v + x) __lowerCAmelCase = int(z - (2 * c)) __lowerCAmelCase = round(w % 7) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError('''The date was evaluated incorrectly. Contact developer.''') # Response __lowerCAmelCase = F"""Your date {date_input}, is a {days[str(lowerCamelCase)]}!""" return response if __name__ == "__main__": import doctest doctest.testmod() _UpperCAmelCase : List[str] = argparse.ArgumentParser( description=( """Find out what day of the week nearly any date is or was. Enter """ """date as a string in the mm-dd-yyyy or mm/dd/yyyy format""" ) ) parser.add_argument( """date_input""", type=str, help="""Date as a string (mm-dd-yyyy or mm/dd/yyyy)""" ) _UpperCAmelCase : Dict = parser.parse_args() zeller(args.date_input)
9
1
def A_ ( A__ ) -> None: a__ : Optional[int] = generate_pascal_triangle(A__ ) for row_idx in range(A__ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=' ' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=' ' ) else: print(triangle[row_idx][col_idx] , end='' ) print() def A_ ( A__ ) -> list[list[int]]: if not isinstance(A__ , A__ ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) a__ : list[list[int]] = [] for current_row_idx in range(A__ ): a__ : Any = populate_current_row(A__ , A__ ) triangle.append(A__ ) return triangle def A_ ( A__ , A__ ) -> list[int]: a__ : Optional[int] = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 a__ , a__ : str = 1, 1 for current_col_idx in range(1 , A__ ): calculate_current_element( A__ , A__ , A__ , A__ ) return current_row def A_ ( A__ , A__ , A__ , A__ , ) -> None: a__ : Dict = triangle[current_row_idx - 1][current_col_idx - 1] a__ : List[str] = triangle[current_row_idx - 1][current_col_idx] a__ : str = above_to_left_elt + above_to_right_elt def A_ ( A__ ) -> list[list[int]]: if not isinstance(A__ , A__ ): raise TypeError('The input value of \'num_rows\' should be \'int\'' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( 'The input value of \'num_rows\' should be greater than or equal to 0' ) a__ : list[list[int]] = [[1]] for row_index in range(1 , A__ ): a__ : str = [0] + result[-1] + [0] a__ : str = row_index + 1 # Calculate the number of distinct elements in a row a__ : Optional[int] = sum(divmod(A__ , 2 ) ) a__ : Dict = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] a__ : Optional[Any] = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() a__ : Any = row_first_half + row_second_half result.append(A__ ) return result def A_ ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(A__ , A__ ) -> None: a__ : List[str] = F'{func.__name__}({value})' a__ : Tuple = timeit(F'__main__.{call}' , setup='import __main__' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'{call:38} -- {timing:.4f} seconds' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(A__ , A__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
99
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCAmelCase = logging.get_logger(__name__) def __UpperCamelCase ( lowercase__ : int, lowercase__ : str ): '''simple docstring''' try: with open(lowercase__, 'rb' ) as flax_state_f: __lowercase =from_bytes(lowercase__, flax_state_f.read() ) except UnpicklingError as e: try: with open(lowercase__ ) as f: if f.read().startswith('version' ): raise OSError( 'You seem to have cloned a repository without having git-lfs installed. Please' ' install git-lfs and run `git lfs install` followed by `git lfs pull` in the' ' folder you cloned.' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F'''Unable to convert {model_file} to Flax deserializable object. ''' ) return load_flax_weights_in_pytorch_model(lowercase__, lowercase__ ) def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : List[str] ): '''simple docstring''' try: import torch # noqa: F401 except ImportError: logger.error( 'Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights __lowercase =flatten_dict(jax.tree_util.tree_map(lambda lowercase__ : x.dtype == jnp.bfloataa, lowercase__ ) ).values() if any(lowercase__ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) __lowercase =jax.tree_util.tree_map( lambda lowercase__ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params, lowercase__ ) __lowercase ='' __lowercase =flatten_dict(lowercase__, sep='.' ) __lowercase =pt_model.state_dict() # keep track of unexpected & missing keys __lowercase =[] __lowercase =set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): __lowercase =flax_key_tuple.split('.' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: __lowercase =flax_key_tuple_array[:-1] + ['weight'] __lowercase =jnp.transpose(lowercase__, (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": __lowercase =flax_key_tuple_array[:-1] + ['weight'] __lowercase =flax_tensor.T elif flax_key_tuple_array[-1] == "scale": __lowercase =flax_key_tuple_array[:-1] + ['weight'] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowercase__ ): __lowercase =( flax_key_tuple_string.replace('_0', '.0' ) .replace('_1', '.1' ) .replace('_2', '.2' ) .replace('_3', '.3' ) .replace('_4', '.4' ) .replace('_5', '.5' ) .replace('_6', '.6' ) .replace('_7', '.7' ) .replace('_8', '.8' ) .replace('_9', '.9' ) ) __lowercase ='.'.join(lowercase__ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'''Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ''' F'''to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) else: # add weight to pytorch dict __lowercase =np.asarray(lowercase__ ) if not isinstance(lowercase__, np.ndarray ) else flax_tensor __lowercase =torch.from_numpy(lowercase__ ) # remove from missing keys missing_keys.remove(lowercase__ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowercase__ ) pt_model.load_state_dict(lowercase__ ) # re-transform missing_keys to list __lowercase =list(lowercase__ ) if len(lowercase__ ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F''' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing''' F''' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture''' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F''' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect''' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) if len(lowercase__ ) > 0: logger.warning( F'''Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly''' F''' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to''' ' use it for predictions and inference.' ) return pt_model
141
0
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position UpperCamelCase = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip UpperCamelCase = concatenate_datasets UpperCamelCase = DownloadConfig UpperCamelCase = DownloadManager UpperCamelCase = DownloadMode UpperCamelCase = DownloadConfig UpperCamelCase = DownloadMode UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
333
import os import re import shutil import sys import tempfile import unittest import black UpperCamelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. UpperCamelCase = ''' \""" Output class for the scheduler\'s step function output. Args: prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the denoising loop. pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): The predicted denoised sample (x_{0}) based on the model output from the current timestep. `pred_original_sample` can be used to preview progress or for guidance. \""" prev_sample: torch.FloatTensor pred_original_sample: Optional[torch.FloatTensor] = None ''' class snake_case_ ( unittest.TestCase ): def __UpperCamelCase ( self : str ) -> List[str]: lowercase__ : str = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir , "schedulers/" ) ) lowercase__ : List[Any] = self.diffusers_dir shutil.copy( os.path.join(lowercase_ , "src/diffusers/schedulers/scheduling_ddpm.py" ) , os.path.join(self.diffusers_dir , "schedulers/scheduling_ddpm.py" ) , ) def __UpperCamelCase ( self : Optional[int] ) -> List[str]: lowercase__ : Dict = "src/diffusers" shutil.rmtree(self.diffusers_dir ) def __UpperCamelCase ( self : Tuple , lowercase_ : Optional[Any] , lowercase_ : Any , lowercase_ : str , lowercase_ : Tuple=None ) -> Tuple: lowercase__ : Optional[int] = comment + F'''\nclass {class_name}(nn.Module):\n''' + class_code if overwrite_result is not None: lowercase__ : Optional[int] = comment + F'''\nclass {class_name}(nn.Module):\n''' + overwrite_result lowercase__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 ) lowercase__ : List[str] = black.format_str(lowercase_ , mode=lowercase_ ) lowercase__ : Optional[int] = os.path.join(self.diffusers_dir , "new_code.py" ) with open(lowercase_ , "w" , newline="\n" ) as f: f.write(lowercase_ ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(lowercase_ ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=lowercase_ ) with open(lowercase_ , "r" ) as f: self.assertTrue(f.read() , lowercase_ ) def __UpperCamelCase ( self : str ) -> Optional[int]: lowercase__ : Optional[Any] = check_copies.find_code_in_diffusers("schedulers.scheduling_ddpm.DDPMSchedulerOutput" ) self.assertEqual(lowercase_ , lowercase_ ) def __UpperCamelCase ( self : int ) -> str: # Base copy consistency self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput" , "DDPMSchedulerOutput" , REFERENCE_CODE + "\n" , ) # With no empty line at the end self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput" , "DDPMSchedulerOutput" , lowercase_ , ) # Copy consistency with rename self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test" , "TestSchedulerOutput" , re.sub("DDPM" , "Test" , lowercase_ ) , ) # Copy consistency with a really long name lowercase__ : Optional[int] = "TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason" self.check_copy_consistency( F'''# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}''' , F'''{long_class_name}SchedulerOutput''' , re.sub("Bert" , lowercase_ , lowercase_ ) , ) # Copy consistency with overwrite self.check_copy_consistency( "# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test" , "TestSchedulerOutput" , lowercase_ , overwrite_result=re.sub("DDPM" , "Test" , lowercase_ ) , )
333
1
"""simple docstring""" import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , UpperCamelCase_ , UpperCamelCase_=13 , UpperCamelCase_=3 , UpperCamelCase_=True , UpperCamelCase_=True , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=2_24 , UpperCamelCase_=10_00 , UpperCamelCase_=[3, 3, 6, 4] , UpperCamelCase_=[48, 56, 1_12, 2_20] , ) -> Dict: __lowercase : str = parent __lowercase : Dict = batch_size __lowercase : Union[str, Any] = num_channels __lowercase : List[str] = is_training __lowercase : int = use_labels __lowercase : List[Any] = hidden_dropout_prob __lowercase : Optional[Any] = attention_probs_dropout_prob __lowercase : Tuple = num_labels __lowercase : Union[str, Any] = image_size __lowercase : Tuple = layer_depths __lowercase : Union[str, Any] = embed_dims def _lowerCamelCase ( self ) -> List[Any]: __lowercase : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase : Dict = None if self.use_labels: __lowercase : List[str] = ids_tensor([self.batch_size] , self.num_labels ) __lowercase : Dict = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self ) -> int: return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act='''gelu''' , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=_SCREAMING_SNAKE_CASE , layer_scale_init_value=1E-5 , ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Dict: __lowercase : Dict = SwiftFormerModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() __lowercase : List[Any] = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> str: __lowercase : Dict = self.num_labels __lowercase : Any = SwiftFormerForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() __lowercase : Union[str, Any] = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __lowercase : Tuple = SwiftFormerForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() __lowercase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase : Any = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self ) -> str: (__lowercase) : Dict = self.prepare_config_and_inputs() __lowercase : Optional[int] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _lowercase , _lowercase , unittest.TestCase ): UpperCamelCase =(SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () UpperCamelCase =( {'''feature-extraction''': SwiftFormerModel, '''image-classification''': SwiftFormerForImageClassification} if is_torch_available() else {} ) UpperCamelCase =False UpperCamelCase =False UpperCamelCase =False UpperCamelCase =False UpperCamelCase =False def _lowerCamelCase ( self ) -> int: __lowercase : List[str] = SwiftFormerModelTester(self ) __lowercase : Dict = ConfigTester( self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def _lowerCamelCase ( self ) -> Optional[Any]: self.config_tester.run_common_tests() @unittest.skip(reason='''SwiftFormer does not use inputs_embeds''' ) def _lowerCamelCase ( self ) -> str: pass def _lowerCamelCase ( self ) -> int: __lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase : Dict = model_class(_SCREAMING_SNAKE_CASE ) __lowercase : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear ) ) def _lowerCamelCase ( self ) -> str: __lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase : str = model_class(_SCREAMING_SNAKE_CASE ) __lowercase : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase : List[Any] = [*signature.parameters.keys()] __lowercase : List[str] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( self ) -> str: __lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( self ) -> Optional[Any]: __lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def _lowerCamelCase ( self ) -> Dict: for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase : List[Any] = SwiftFormerModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) @unittest.skip(reason='''SwiftFormer does not output attentions''' ) def _lowerCamelCase ( self ) -> Any: pass def _lowerCamelCase ( self ) -> str: def check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __lowercase : List[Any] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __lowercase : Union[str, Any] = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) __lowercase : List[str] = outputs.hidden_states __lowercase : str = 8 self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(_SCREAMING_SNAKE_CASE ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase : Any = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase : Any = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( self ) -> Dict: def _config_zero_init(UpperCamelCase_ ): __lowercase : Union[str, Any] = copy.deepcopy(_SCREAMING_SNAKE_CASE ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 1E-10 ) if isinstance(getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ): __lowercase : Optional[int] = _config_zero_init(getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) setattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return configs_no_init __lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __lowercase : List[Any] = _config_zero_init(_SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: __lowercase : Dict = model_class(config=_SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _lowerCamelCase ( self ) -> Any: pass def __UpperCAmelCase ( ): __lowercase : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self ) -> Tuple: return ViTImageProcessor.from_pretrained('''MBZUAI/swiftformer-xs''' ) if is_vision_available() else None @slow def _lowerCamelCase ( self ) -> Any: __lowercase : List[str] = SwiftFormerForImageClassification.from_pretrained('''MBZUAI/swiftformer-xs''' ).to(_SCREAMING_SNAKE_CASE ) __lowercase : Optional[Any] = self.default_image_processor __lowercase : int = prepare_img() __lowercase : Any = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): __lowercase : Dict = model(**_SCREAMING_SNAKE_CASE ) # verify the logits __lowercase : str = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) __lowercase : Optional[Any] = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
249
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch lowerCamelCase = logging.get_logger(__name__) class _a ( _lowercase): _a : Optional[Any] = ['''pixel_values'''] def __init__( self : List[Any] , _SCREAMING_SNAKE_CASE : bool = True , _SCREAMING_SNAKE_CASE : Optional[Dict[str, int]] = None , _SCREAMING_SNAKE_CASE : PILImageResampling = PILImageResampling.BILINEAR , _SCREAMING_SNAKE_CASE : bool = True , _SCREAMING_SNAKE_CASE : Dict[str, int] = None , _SCREAMING_SNAKE_CASE : bool = True , _SCREAMING_SNAKE_CASE : Union[int, float] = 1 / 255 , _SCREAMING_SNAKE_CASE : bool = True , _SCREAMING_SNAKE_CASE : Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE : Optional[Union[float, List[float]]] = None , **_SCREAMING_SNAKE_CASE : int , )-> None: super().__init__(**_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = size if size is not None else {'''shortest_edge''': 256} lowerCAmelCase__ : Tuple = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} lowerCAmelCase__ : Optional[Any] = get_size_dict(_SCREAMING_SNAKE_CASE , param_name='''crop_size''' ) lowerCAmelCase__ : List[str] = do_resize lowerCAmelCase__ : Optional[Any] = size lowerCAmelCase__ : Any = resample lowerCAmelCase__ : str = do_center_crop lowerCAmelCase__ : Dict = crop_size lowerCAmelCase__ : str = do_rescale lowerCAmelCase__ : List[str] = rescale_factor lowerCAmelCase__ : int = do_normalize lowerCAmelCase__ : Dict = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase__ : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : Dict[str, int] , _SCREAMING_SNAKE_CASE : PILImageResampling = PILImageResampling.BICUBIC , _SCREAMING_SNAKE_CASE : Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE : Dict , )-> np.ndarray: lowerCAmelCase__ : str = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE ) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) lowerCAmelCase__ : List[str] = get_resize_output_image_size(_SCREAMING_SNAKE_CASE , size=size['''shortest_edge'''] , default_to_square=_SCREAMING_SNAKE_CASE ) return resize(_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : Dict[str, int] , _SCREAMING_SNAKE_CASE : Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE : List[str] , )-> np.ndarray: lowerCAmelCase__ : Dict = get_size_dict(_SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}' ) return center_crop(_SCREAMING_SNAKE_CASE , size=(size['''height'''], size['''width''']) , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE : Optional[int] )-> np.ndarray: return rescale(_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : Union[float, List[float]] , _SCREAMING_SNAKE_CASE : Union[float, List[float]] , _SCREAMING_SNAKE_CASE : Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE : str , )-> np.ndarray: return normalize(_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Any , _SCREAMING_SNAKE_CASE : ImageInput , _SCREAMING_SNAKE_CASE : Optional[bool] = None , _SCREAMING_SNAKE_CASE : Dict[str, int] = None , _SCREAMING_SNAKE_CASE : PILImageResampling = None , _SCREAMING_SNAKE_CASE : bool = None , _SCREAMING_SNAKE_CASE : Dict[str, int] = None , _SCREAMING_SNAKE_CASE : Optional[bool] = None , _SCREAMING_SNAKE_CASE : Optional[float] = None , _SCREAMING_SNAKE_CASE : Optional[bool] = None , _SCREAMING_SNAKE_CASE : Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE : Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE : Optional[Union[str, TensorType]] = None , _SCREAMING_SNAKE_CASE : Union[str, ChannelDimension] = ChannelDimension.FIRST , **_SCREAMING_SNAKE_CASE : Tuple , )-> Optional[Any]: lowerCAmelCase__ : List[str] = do_resize if do_resize is not None else self.do_resize lowerCAmelCase__ : List[str] = size if size is not None else self.size lowerCAmelCase__ : Any = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = resample if resample is not None else self.resample lowerCAmelCase__ : Dict = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCAmelCase__ : Dict = crop_size if crop_size is not None else self.crop_size lowerCAmelCase__ : Any = get_size_dict(_SCREAMING_SNAKE_CASE , param_name='''crop_size''' ) lowerCAmelCase__ : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase__ : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase__ : str = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase__ : List[Any] = image_mean if image_mean is not None else self.image_mean lowerCAmelCase__ : List[str] = image_std if image_std is not None else self.image_std lowerCAmelCase__ : Optional[int] = make_list_of_images(_SCREAMING_SNAKE_CASE ) if not valid_images(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. lowerCAmelCase__ : List[Any] = [to_numpy_array(_SCREAMING_SNAKE_CASE ) for image in images] if do_resize: lowerCAmelCase__ : Dict = [self.resize(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE ) for image in images] if do_center_crop: lowerCAmelCase__ : Dict = [self.center_crop(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE ) for image in images] if do_rescale: lowerCAmelCase__ : List[Any] = [self.rescale(image=_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE ) for image in images] if do_normalize: lowerCAmelCase__ : Tuple = [self.normalize(image=_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE ) for image in images] lowerCAmelCase__ : Dict = [to_channel_dimension_format(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for image in images] lowerCAmelCase__ : Dict = {'''pixel_values''': images} return BatchFeature(data=_SCREAMING_SNAKE_CASE , tensor_type=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[Tuple] = None )-> List[Any]: lowerCAmelCase__ : Union[str, Any] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(_SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : Tuple = target_sizes.numpy() lowerCAmelCase__ : Tuple = [] for idx in range(len(_SCREAMING_SNAKE_CASE ) ): lowerCAmelCase__ : int = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_SCREAMING_SNAKE_CASE ) else: lowerCAmelCase__ : Any = logits.argmax(dim=1 ) lowerCAmelCase__ : Dict = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
131
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''facebook/levit-128S''': '''https://huggingface.co/facebook/levit-128S/resolve/main/config.json''', # See all LeViT models at https://huggingface.co/models?filter=levit } class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = '''levit''' def __init__( self : str , __UpperCAmelCase : int=224 , __UpperCAmelCase : Optional[Any]=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : Any=1 , __UpperCAmelCase : int=16 , __UpperCAmelCase : Any=[128, 256, 384] , __UpperCAmelCase : Optional[Any]=[4, 8, 12] , __UpperCAmelCase : Dict=[4, 4, 4] , __UpperCAmelCase : Union[str, Any]=[16, 16, 16] , __UpperCAmelCase : List[Any]=0 , __UpperCAmelCase : str=[2, 2, 2] , __UpperCAmelCase : Optional[Any]=[2, 2, 2] , __UpperCAmelCase : int=0.02 , **__UpperCAmelCase : Dict , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) _A = image_size _A = num_channels _A = kernel_size _A = stride _A = padding _A = hidden_sizes _A = num_attention_heads _A = depths _A = key_dim _A = drop_path_rate _A = patch_size _A = attention_ratio _A = mlp_ratio _A = initializer_range _A = [ ["Subsample", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["Subsample", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = version.parse('''1.11''' ) @property def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase ( self : Any ): '''simple docstring''' return 1E-4
174
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCamelCase_ = { '''configuration_trocr''': ['''TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrOCRConfig'''], '''processing_trocr''': ['''TrOCRProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TROCR_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrOCRForCausalLM''', '''TrOCRPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
174
1
"""simple docstring""" from math import isqrt def _A ( lowercase ): """simple docstring""" a =[True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowercase , lowercase ): a =False return [i for i in range(2 , lowercase ) if is_prime[i]] def _A ( lowercase = 10**8 ): """simple docstring""" a =calculate_prime_numbers(max_number // 2 ) a =0 a =0 a =len(lowercase ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(F'{solution() = }')
81
"""simple docstring""" import requests from bsa import BeautifulSoup def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = BeautifulSoup(requests.get(lowerCamelCase , params=lowerCamelCase ).content , 'html.parser' ) UpperCAmelCase__ = soup.find('div' , attrs={'class': 'gs_ri'} ) UpperCAmelCase__ = div.find('div' , attrs={'class': 'gs_fl'} ).find_all('a' ) return anchors[2].get_text() if __name__ == "__main__": lowerCAmelCase__ : Optional[int] = { 'title': ( 'Precisely geometry controlled microsupercapacitors for ultrahigh areal ' 'capacitance, volumetric capacitance, and energy density' ), 'journal': 'Chem. Mater.', 'volume': 30, 'pages': '3979-3990', 'year': 2_018, 'hl': 'en', } print(get_citation('https://scholar.google.com/scholar_lookup', params=params))
98
0
import comet # From: unbabel-comet import torch import datasets SCREAMING_SNAKE_CASE :str = datasets.logging.get_logger(__name__) SCREAMING_SNAKE_CASE :List[Any] = '\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task},\n booktitle = {Proceedings of the Fifth Conference on Machine Translation},\n month = {November},\n year = {2020},\n address = {Online},\n publisher = {Association for Computational Linguistics},\n pages = {909--918},\n}\n@inproceedings{rei-etal-2020-comet,\n title = \"{COMET}: A Neural Framework for {MT} Evaluation\",\n author = \"Rei, Ricardo and\n Stewart, Craig and\n Farinha, Ana C and\n Lavie, Alon\",\n booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\",\n month = nov,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\",\n pages = \"2685--2702\",\n}\n' SCREAMING_SNAKE_CASE :Tuple = '\\nCrosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM).\nWith the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition.\n\nSee the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information.\n' SCREAMING_SNAKE_CASE :Union[str, Any] = '\nCOMET score.\n\nArgs:\n\n`sources` (list of str): Source sentences\n`predictions` (list of str): candidate translations\n`references` (list of str): reference translations\n`cuda` (bool): If set to True, runs COMET using GPU\n`show_progress` (bool): Shows progress\n`model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None.\n\nReturns:\n `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`.\n `scores`: List of scores.\n\nExamples:\n\n >>> comet_metric = datasets.load_metric(\'comet\')\n >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use\n >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"]\n >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"]\n >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"]\n >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source)\n >>> print([round(v, 2) for v in results[\"scores\"]])\n [0.19, 0.92]\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def UpperCamelCase_ ( self : List[Any] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,homepage="https://unbabel.github.io/COMET/html/index.html" ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { "sources": datasets.Value("string" ,id="sequence" ), "predictions": datasets.Value("string" ,id="sequence" ), "references": datasets.Value("string" ,id="sequence" ), } ) ,codebase_urls=["https://github.com/Unbabel/COMET"] ,reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] ,) def UpperCamelCase_ ( self : Tuple ,A : Any ): if self.config_name == "default": __A = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da" ) ) else: __A = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def UpperCamelCase_ ( self : Optional[int] ,A : Optional[Any] ,A : Optional[Any] ,A : int ,A : Optional[int]=None ,A : List[str]=False ): if gpus is None: __A = 1 if torch.cuda.is_available() else 0 __A = {"src": sources, "mt": predictions, "ref": references} __A = [dict(zip(UpperCamelCase__ ,UpperCamelCase__ ) ) for t in zip(*data.values() )] __A = self.scorer.predict(UpperCamelCase__ ,gpus=UpperCamelCase__ ,progress_bar=UpperCamelCase__ ) return {"mean_score": mean_score, "scores": scores}
370
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split SCREAMING_SNAKE_CASE :Tuple = datasets.load_iris() SCREAMING_SNAKE_CASE :Dict = np.array(data['data']) SCREAMING_SNAKE_CASE :Optional[int] = np.array(data['target']) SCREAMING_SNAKE_CASE :List[str] = data['target_names'] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE :List[Any] = train_test_split(X, y) def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" return np.linalg.norm(np.array(a_ ) - np.array(a_ ) ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_=5 ) -> Dict: """simple docstring""" __A = zip(a_ , a_ ) # List of distances of all points from the point to be classified __A = [] for data_point in data: __A = euclidean_distance(data_point[0] , a_ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. __A = [i[1] for i in sorted(a_ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified __A = Counter(a_ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
124
0
'''simple docstring''' import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def __lowercase ( __lowercase ) -> Dict: '''simple docstring''' if isinstance(__lowercase , collections.abc.Iterable ): return x return (x, x) @require_flax class _UpperCAmelCase : """simple docstring""" def lowerCAmelCase ( self : int , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Dict ): '''simple docstring''' pass def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' pass def lowerCAmelCase ( self : List[str] ): '''simple docstring''' pass def lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : float ): '''simple docstring''' _A = np.abs((a - b) ).max() self.assertLessEqual(__UpperCAmelCase , __UpperCAmelCase , f'''Difference between torch and flax is {diff} (>= {tol}).''' ) def lowerCAmelCase ( self : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int=None , **__UpperCAmelCase : Optional[int] ): '''simple docstring''' _A = VisionTextDualEncoderConfig.from_vision_text_configs(__UpperCAmelCase , __UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel(__UpperCAmelCase ) _A = model(input_ids=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def lowerCAmelCase ( self : str , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[Any]=None , **__UpperCAmelCase : Dict ): '''simple docstring''' _A , _A = self.get_vision_text_model(__UpperCAmelCase , __UpperCAmelCase ) _A = {"vision_model": vision_model, "text_model": text_model} _A = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__UpperCAmelCase ) _A = model(input_ids=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Dict , __UpperCAmelCase : int , __UpperCAmelCase : Any , __UpperCAmelCase : Optional[Any]=None , **__UpperCAmelCase : str ): '''simple docstring''' _A , _A = self.get_vision_text_model(__UpperCAmelCase , __UpperCAmelCase ) _A = {"vision_model": vision_model, "text_model": text_model} _A = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__UpperCAmelCase ) _A = model(input_ids=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase ) _A = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel.from_pretrained(__UpperCAmelCase ) _A = model(input_ids=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase ) _A = after_output[0] _A = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__UpperCAmelCase , 1E-3 ) def lowerCAmelCase ( self : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str=None , **__UpperCAmelCase : List[Any] ): '''simple docstring''' _A , _A = self.get_vision_text_model(__UpperCAmelCase , __UpperCAmelCase ) _A = {"vision_model": vision_model, "text_model": text_model} _A = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__UpperCAmelCase ) _A = model( input_ids=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase ) _A = output.vision_model_output.attentions self.assertEqual(len(__UpperCAmelCase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _A = to_atuple(vision_model.config.image_size ) _A = to_atuple(vision_model.config.patch_size ) _A = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _A = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _A = output.text_model_output.attentions self.assertEqual(len(__UpperCAmelCase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : int ): '''simple docstring''' pt_model.to(__UpperCAmelCase ) pt_model.eval() # prepare inputs _A = inputs_dict _A = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): _A = pt_model(**__UpperCAmelCase ).to_tuple() _A = fx_model(**__UpperCAmelCase ).to_tuple() self.assertEqual(len(__UpperCAmelCase ) , len(__UpperCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(__UpperCAmelCase , pt_output.numpy() , 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(__UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel.from_pretrained(__UpperCAmelCase , from_pt=__UpperCAmelCase ) _A = fx_model_loaded(**__UpperCAmelCase ).to_tuple() self.assertEqual(len(__UpperCAmelCase ) , len(__UpperCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(__UpperCAmelCase , pt_output.numpy() , 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(__UpperCAmelCase ) _A = VisionTextDualEncoderModel.from_pretrained(__UpperCAmelCase , from_flax=__UpperCAmelCase ) pt_model_loaded.to(__UpperCAmelCase ) pt_model_loaded.eval() with torch.no_grad(): _A = pt_model_loaded(**__UpperCAmelCase ).to_tuple() self.assertEqual(len(__UpperCAmelCase ) , len(__UpperCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(__UpperCAmelCase , pt_output_loaded.numpy() , 4E-2 ) def lowerCAmelCase ( self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : str , __UpperCAmelCase : str ): '''simple docstring''' _A = VisionTextDualEncoderConfig.from_vision_text_configs(__UpperCAmelCase , __UpperCAmelCase ) _A = VisionTextDualEncoderModel(__UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel(__UpperCAmelCase ) _A = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , __UpperCAmelCase ) _A = fx_state self.check_pt_flax_equivalence(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowerCAmelCase ( self : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[str] ): '''simple docstring''' _A = VisionTextDualEncoderConfig.from_vision_text_configs(__UpperCAmelCase , __UpperCAmelCase ) _A = VisionTextDualEncoderModel(__UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel(__UpperCAmelCase ) _A = load_flax_weights_in_pytorch_model(__UpperCAmelCase , fx_model.params ) self.check_pt_flax_equivalence(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' _A = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**__UpperCAmelCase ) def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' _A = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**__UpperCAmelCase ) def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' _A = self.prepare_config_and_inputs() self.check_save_load(**__UpperCAmelCase ) def lowerCAmelCase ( self : Tuple ): '''simple docstring''' _A = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**__UpperCAmelCase ) @is_pt_flax_cross_test def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' _A = self.prepare_config_and_inputs() _A = config_inputs_dict.pop("vision_config" ) _A = config_inputs_dict.pop("text_config" ) _A = config_inputs_dict self.check_equivalence_pt_to_flax(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) self.check_equivalence_flax_to_pt(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) @slow def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' _A , _A = self.get_pretrained_model_and_inputs() _A = model_a(**__UpperCAmelCase ) _A = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(__UpperCAmelCase ) _A = FlaxVisionTextDualEncoderModel.from_pretrained(__UpperCAmelCase ) _A = model_a(**__UpperCAmelCase ) _A = after_outputs[0] _A = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__UpperCAmelCase , 1E-5 ) @require_flax class _UpperCAmelCase ( snake_case_ , unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' _A = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=__UpperCAmelCase , text_from_pt=__UpperCAmelCase , ) _A = 13 _A = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _A = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _A = random_attention_mask([batch_size, 4] ) _A = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any] ): '''simple docstring''' _A = FlaxViTModel(__UpperCAmelCase ) _A = FlaxBertModel(__UpperCAmelCase ) return vision_model, text_model def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' _A = FlaxViTModelTester(self ) _A = FlaxBertModelTester(self ) _A = vit_model_tester.prepare_config_and_inputs() _A = bert_model_tester.prepare_config_and_inputs() _A , _A = vision_config_and_inputs _A , _A , _A , _A = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class _UpperCAmelCase ( snake_case_ , unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : int ): '''simple docstring''' _A = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=__UpperCAmelCase , text_from_pt=__UpperCAmelCase , ) _A = 13 _A = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _A = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _A = random_attention_mask([batch_size, 4] ) _A = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase ( self : Tuple , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict ): '''simple docstring''' _A = FlaxCLIPVisionModel(__UpperCAmelCase ) _A = FlaxBertModel(__UpperCAmelCase ) return vision_model, text_model def lowerCAmelCase ( self : Tuple ): '''simple docstring''' _A = FlaxCLIPVisionModelTester(self ) _A = FlaxBertModelTester(self ) _A = clip_model_tester.prepare_config_and_inputs() _A = bert_model_tester.prepare_config_and_inputs() _A , _A = vision_config_and_inputs _A , _A , _A , _A = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowerCAmelCase ( self : int ): '''simple docstring''' _A = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0 ) _A = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) _A = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) _A = processor( text=["una foto di un gatto", "una foto di un cane"] , images=__UpperCAmelCase , padding=__UpperCAmelCase , return_tensors="np" ) _A = model(**__UpperCAmelCase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _A = np.array([[1.2284727, 0.3104122]] ) self.assertTrue(np.allclose(outputs.logits_per_image , __UpperCAmelCase , atol=1E-3 ) )
79
from __future__ import annotations from typing import Any class _a : def __init__(self, SCREAMING_SNAKE_CASE_ = 6 ) -> None: UpperCAmelCase_: Node | None = None UpperCAmelCase_: Node | None = None self.create_linked_list(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCAmelCase_: Optional[Any] = Node() UpperCAmelCase_: Optional[Any] = current_node UpperCAmelCase_: List[str] = current_node UpperCAmelCase_: List[Any] = current_node for _ in range(1, SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Optional[int] = Node() UpperCAmelCase_: Dict = current_node UpperCAmelCase_: Any = previous_node UpperCAmelCase_: Tuple = current_node UpperCAmelCase_: Optional[Any] = self.front UpperCAmelCase_: Any = previous_node def __snake_case (self ) -> bool: return ( self.front == self.rear and self.front is not None and self.front.data is None ) def __snake_case (self ) -> Any | None: self.check_can_perform_operation() return self.front.data if self.front else None def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> None: if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase_: Optional[int] = self.rear.next if self.rear: UpperCAmelCase_: Any = data def __snake_case (self ) -> Any: self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase_: Union[str, Any] = self.front.data UpperCAmelCase_: Any = None return data UpperCAmelCase_: str = self.front UpperCAmelCase_: Union[str, Any] = old_front.next UpperCAmelCase_: int = old_front.data UpperCAmelCase_: Any = None return data def __snake_case (self ) -> None: if self.is_empty(): raise Exception("""Empty Queue""" ) def __snake_case (self ) -> None: if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class _a : def __init__(self ) -> None: UpperCAmelCase_: Any | None = None UpperCAmelCase_: Node | None = None UpperCAmelCase_: Node | None = None if __name__ == "__main__": import doctest doctest.testmod()
147
0
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss UpperCamelCase : List[str] = pytest.mark.integration @require_faiss class UpperCamelCase ( lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any]): """simple docstring""" a : Tuple = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(__lowerCamelCase) for x in np.arange(3_0).tolist()]}) return dset def SCREAMING_SNAKE_CASE_ ( self : str): """simple docstring""" import faiss a : Dataset = self._create_dummy_dataset() a : Dict = dset.map( lambda UpperCAmelCase_ , UpperCAmelCase_: {"vecs": i * np.ones(5 , dtype=np.floataa)} , with_indices=__lowerCamelCase , keep_in_memory=__lowerCamelCase) a : str = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT) a : str = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa)) self.assertEqual(examples['filename'][0] , 'my_name-train_29') dset.drop_index('vecs') def SCREAMING_SNAKE_CASE_ ( self : str): """simple docstring""" import faiss a : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5)) * np.arange(3_0).reshape(-1 , 1) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) a : Union[str, Any] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa)) self.assertEqual(examples['filename'][0] , 'my_name-train_29') def SCREAMING_SNAKE_CASE_ ( self : Optional[Any]): """simple docstring""" import faiss a : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5)) * np.arange(3_0).reshape(-1 , 1) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=__lowerCamelCase) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name) dset.load_faiss_index('vecs2' , tmp_file.name) os.unlink(tmp_file.name) a : List[Any] = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa)) self.assertEqual(examples['filename'][0] , 'my_name-train_29') def SCREAMING_SNAKE_CASE_ ( self : Optional[Any]): """simple docstring""" a : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5)) * np.arange(3_0).reshape(-1 , 1) , index_name='vecs') dset.drop_index('vecs') self.assertRaises(__lowerCamelCase , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa))) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any]): """simple docstring""" from elasticsearch import Elasticsearch a : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search') as mocked_search, patch( 'elasticsearch.client.IndicesClient.create') as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk') as mocked_bulk: a : Optional[Any] = {'''acknowledged''': True} mocked_bulk.return_value([(True, None)] * 3_0) a : List[str] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 2_9}]}} a : Optional[int] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=__lowerCamelCase) a : Tuple = dset.get_nearest_examples('filename' , 'my_name-train_29') self.assertEqual(examples['filename'][0] , 'my_name-train_29') @require_faiss class UpperCamelCase ( lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self : Any): """simple docstring""" import faiss a : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa)) self.assertIsNotNone(index.faiss_index) self.assertEqual(index.faiss_index.ntotal , 5) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa)) self.assertEqual(index.faiss_index.ntotal , 1_0) # single query a : Union[str, Any] = np.zeros(5 , dtype=np.floataa) a : str = 1 a : Any = index.search(__lowerCamelCase) self.assertRaises(__lowerCamelCase , index.search , query.reshape(-1 , 1)) self.assertGreater(scores[0] , 0) self.assertEqual(indices[0] , 1) # batched queries a : Optional[Any] = np.eye(5 , dtype=np.floataa)[::-1] a : List[Any] = index.search_batch(__lowerCamelCase) self.assertRaises(__lowerCamelCase , index.search_batch , queries[0]) a : Union[str, Any] = [scores[0] for scores in total_scores] a : Any = [indices[0] for indices in total_indices] self.assertGreater(np.min(__lowerCamelCase) , 0) self.assertListEqual([4, 3, 2, 1, 0] , __lowerCamelCase) def SCREAMING_SNAKE_CASE_ ( self : List[str]): """simple docstring""" import faiss a : List[str] = FaissIndex(string_factory='Flat') index.add_vectors(np.eye(5 , dtype=np.floataa)) self.assertIsInstance(index.faiss_index , faiss.IndexFlat) a : Union[str, Any] = FaissIndex(string_factory='LSH') index.add_vectors(np.eye(5 , dtype=np.floataa)) self.assertIsInstance(index.faiss_index , faiss.IndexLSH) with self.assertRaises(__lowerCamelCase): a : int = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5)) def SCREAMING_SNAKE_CASE_ ( self : Optional[int]): """simple docstring""" import faiss a : Any = faiss.IndexFlat(5) a : Tuple = FaissIndex(custom_index=__lowerCamelCase) index.add_vectors(np.eye(5 , dtype=np.floataa)) self.assertIsInstance(index.faiss_index , faiss.IndexFlat) def SCREAMING_SNAKE_CASE_ ( self : List[str]): """simple docstring""" import faiss a : Dict = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT) index.add_vectors(np.eye(5 , dtype=np.floataa)) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=__lowerCamelCase) as tmp_file: index.save(tmp_file.name) a : Any = FaissIndex.load(tmp_file.name) os.unlink(tmp_file.name) a : List[Any] = np.zeros(5 , dtype=np.floataa) a : Any = 1 a : List[str] = index.search(__lowerCamelCase) self.assertGreater(scores[0] , 0) self.assertEqual(indices[0] , 1) @require_faiss def SCREAMING_SNAKE_CASE__ ( snake_case : str ) -> List[Any]: """simple docstring""" import faiss a : List[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) a : Union[str, Any] = '''index.faiss''' a : List[Any] = F"""mock://{index_name}""" index.save(lowerCamelCase_ , storage_options=mockfs.storage_options ) a : Optional[Any] = FaissIndex.load(lowerCamelCase_ , storage_options=mockfs.storage_options ) a : Any = np.zeros(5 , dtype=np.floataa ) a : Optional[Any] = 1 a : Tuple = index.search(lowerCamelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class UpperCamelCase ( lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self : int): """simple docstring""" from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search') as mocked_search, patch( 'elasticsearch.client.IndicesClient.create') as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk') as mocked_bulk: a : Any = Elasticsearch() a : Tuple = {'''acknowledged''': True} a : Union[str, Any] = ElasticSearchIndex(es_client=__lowerCamelCase) mocked_bulk.return_value([(True, None)] * 3) index.add_documents(['foo', 'bar', 'foobar']) # single query a : Optional[int] = '''foo''' a : Optional[int] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 0}]}} a : Optional[Any] = index.search(__lowerCamelCase) self.assertEqual(scores[0] , 1) self.assertEqual(indices[0] , 0) # single query with timeout a : Optional[Any] = '''foo''' a : Dict = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 0}]}} a : Tuple = index.search(__lowerCamelCase , request_timeout=3_0) self.assertEqual(scores[0] , 1) self.assertEqual(indices[0] , 0) # batched queries a : List[str] = ['''foo''', '''bar''', '''foobar'''] a : List[Any] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 1}]}} a : Dict = index.search_batch(__lowerCamelCase) a : Dict = [scores[0] for scores in total_scores] a : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(__lowerCamelCase) , 0) self.assertListEqual([1, 1, 1] , __lowerCamelCase) # batched queries with timeout a : Optional[Any] = ['''foo''', '''bar''', '''foobar'''] a : Any = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 1}]}} a : Dict = index.search_batch(__lowerCamelCase , request_timeout=3_0) a : List[Any] = [scores[0] for scores in total_scores] a : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(__lowerCamelCase) , 0) self.assertListEqual([1, 1, 1] , __lowerCamelCase)
356
'''simple docstring''' import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class UpperCamelCase ( a_ , unittest.TestCase ): """simple docstring""" A : Union[str, Any] = CTRLTokenizer A : List[Any] = False A : Optional[Any] = False def SCREAMING_SNAKE_CASE_ ( self : Dict): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt a : Dict = ['adapt', 're@@', 'a@@', 'apt', 'c@@', 't', '<unk>'] a : Tuple = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_)))) a : Any = ['#version: 0.2', 'a p', 'ap t</w>', 'r e', 'a d', 'ad apt</w>', ''] a : List[Any] = {'unk_token': '<unk>'} a : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(UpperCAmelCase_) + '\n') with open(self.merges_file , 'w' , encoding='utf-8') as fp: fp.write('\n'.join(UpperCAmelCase_)) def SCREAMING_SNAKE_CASE_ ( self : Tuple , **UpperCAmelCase_ : Dict): """simple docstring""" kwargs.update(self.special_tokens_map) return CTRLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : Tuple , UpperCAmelCase_ : Any): """simple docstring""" a : List[str] = 'adapt react readapt apt' a : int = 'adapt react readapt apt' return input_text, output_text def SCREAMING_SNAKE_CASE_ ( self : Optional[int]): """simple docstring""" a : int = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) a : str = 'adapt react readapt apt' a : Optional[Any] = 'adapt re@@ a@@ c@@ t re@@ adapt apt'.split() a : List[Any] = tokenizer.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) a : Union[str, Any] = tokens + [tokenizer.unk_token] a : Any = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_) , UpperCAmelCase_)
345
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _snake_case = logging.get_logger(__name__) _snake_case = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} _snake_case = { "vocab_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/vocab.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/vocab.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/vocab.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/vocab.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/vocab.json", }, "merges_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/merges.txt", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/merges.txt", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/merges.txt", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/merges.txt", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/merges.txt", }, "tokenizer_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/tokenizer.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/tokenizer.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/tokenizer.json", }, } _snake_case = { "gpt2": 1024, "gpt2-medium": 1024, "gpt2-large": 1024, "gpt2-xl": 1024, "distilgpt2": 1024, } class UpperCAmelCase_ ( a): lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ['input_ids', 'attention_mask'] lowerCamelCase__ = GPTaTokenizer def __init__( self, __a=None, __a=None, __a=None, __a="<|endoftext|>", __a="<|endoftext|>", __a="<|endoftext|>", __a=False, **__a, ): '''simple docstring''' super().__init__( __a, __a, tokenizer_file=__a, unk_token=__a, bos_token=__a, eos_token=__a, add_prefix_space=__a, **__a, ) _lowerCAmelCase : Tuple = kwargs.pop("add_bos_token", __a) _lowerCAmelCase : Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space", __a) != add_prefix_space: _lowerCAmelCase : Optional[int] = getattr(__a, pre_tok_state.pop("type")) _lowerCAmelCase : str = add_prefix_space _lowerCAmelCase : Tuple = pre_tok_class(**__a) _lowerCAmelCase : int = add_prefix_space def snake_case__ ( self, *__a, **__a): '''simple docstring''' _lowerCAmelCase : List[str] = kwargs.get("is_split_into_words", __a) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__a, **__a) def snake_case__ ( self, *__a, **__a): '''simple docstring''' _lowerCAmelCase : str = kwargs.get("is_split_into_words", __a) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*__a, **__a) def snake_case__ ( self, __a, __a = None): '''simple docstring''' _lowerCAmelCase : int = self._tokenizer.model.save(__a, name=__a) return tuple(__a) def snake_case__ ( self, __a): '''simple docstring''' _lowerCAmelCase : int = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__a, add_special_tokens=__a) + [self.eos_token_id]) if len(__a) > self.model_max_length: _lowerCAmelCase : Any = input_ids[-self.model_max_length :] return input_ids
36
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : List[Any] = "huggingface/label-files" _lowerCAmelCase : int = "imagenet-1k-id2label.json" _lowerCAmelCase : Tuple = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type="dataset" ) , "r" ) ) _lowerCAmelCase : Tuple = {int(_lowerCamelCase ): v for k, v in idalabel.items()} _lowerCAmelCase : Union[str, Any] = {v: k for k, v in idalabel.items()} _lowerCAmelCase : Tuple = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" _lowerCAmelCase : Optional[int] = BitConfig( conv_layer=_lowerCamelCase , num_labels=1_000 , idalabel=_lowerCamelCase , labelaid=_lowerCamelCase , ) return config def A ( _lowerCamelCase ): '''simple docstring''' if "stem.conv" in name: _lowerCAmelCase : List[str] = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: _lowerCAmelCase : Any = name.replace("blocks" , "layers" ) if "head.fc" in name: _lowerCAmelCase : Optional[Any] = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): _lowerCAmelCase : Any = "bit." + name if "bit" not in name and "classifier" not in name: _lowerCAmelCase : Dict = "bit.encoder." + name return name def A ( ): '''simple docstring''' _lowerCAmelCase : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" _lowerCAmelCase : Optional[int] = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return im @torch.no_grad() def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ): '''simple docstring''' _lowerCAmelCase : Dict = get_config(_lowerCamelCase ) # load original model from timm _lowerCAmelCase : int = create_model(_lowerCamelCase , pretrained=_lowerCamelCase ) timm_model.eval() # load state_dict of original model _lowerCAmelCase : Any = timm_model.state_dict() for key in state_dict.copy().keys(): _lowerCAmelCase : Dict = state_dict.pop(_lowerCamelCase ) _lowerCAmelCase : Tuple = val.squeeze() if "head" in key else val # load HuggingFace model _lowerCAmelCase : Optional[Any] = BitForImageClassification(_lowerCamelCase ) model.eval() model.load_state_dict(_lowerCamelCase ) # create image processor _lowerCAmelCase : Dict = create_transform(**resolve_data_config({} , model=_lowerCamelCase ) ) _lowerCAmelCase : Optional[int] = transform.transforms _lowerCAmelCase : Tuple = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } _lowerCAmelCase : Tuple = BitImageProcessor( do_resize=_lowerCamelCase , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_lowerCamelCase , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_lowerCamelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Any = transform(_lowerCamelCase ).unsqueeze(0 ) _lowerCAmelCase : Optional[int] = processor(_lowerCamelCase , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_lowerCamelCase , _lowerCamelCase ) # verify logits with torch.no_grad(): _lowerCAmelCase : Tuple = model(_lowerCamelCase ) _lowerCAmelCase : str = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) _lowerCAmelCase : Union[str, Any] = timm_model(_lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCamelCase , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) print(F"Saving model {model_name} and processor to {pytorch_dump_folder_path}" ) model.save_pretrained(_lowerCamelCase ) processor.save_pretrained(_lowerCamelCase ) if push_to_hub: print(F"Pushing model {model_name} and processor to the hub" ) model.push_to_hub(F"ybelkada/{model_name}" ) processor.push_to_hub(F"ybelkada/{model_name}" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="resnetv2_50x1_bitm", type=str, help="Name of the BiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model to the hub.", ) _snake_case = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
36
1
"""simple docstring""" import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path __A = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def _lowerCamelCase(__UpperCamelCase=True ) -> Optional[Any]: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__magic_name__ ) ) class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' lowerCamelCase = None lowerCamelCase = None def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: with TemporaryDirectory() as tmp_dir: _lowerCAmelCase =dataset_module_factory(__UpperCAmelCase , cache_dir=__UpperCAmelCase ) _lowerCAmelCase =import_main_class(dataset_module.module_path , dataset=__UpperCAmelCase ) _lowerCAmelCase =builder_cls( cache_dir=__UpperCAmelCase , config_name=__UpperCAmelCase , hash=dataset_module.hash , ) _lowerCAmelCase ="""/""".join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=__UpperCAmelCase ).replace(os.sep , """/""" ), config.DATASET_INFO_FILENAME, ] ) _lowerCAmelCase =cached_path(__UpperCAmelCase , cache_dir=__UpperCAmelCase ) self.assertTrue(os.path.exists(__UpperCAmelCase ) ) @pytest.mark.integration def _lowerCamelCase(__UpperCamelCase ) -> Optional[Any]: _lowerCAmelCase =tmp_path_factory.mktemp("""test_hf_gcp""" ) / """test_wikipedia_simple""" _lowerCAmelCase =dataset_module_factory("""wikipedia""" , cache_dir=__UpperCamelCase ) _lowerCAmelCase =import_main_class(dataset_module.module_path ) _lowerCAmelCase =builder_cls( cache_dir=__UpperCamelCase , config_name="""20220301.frr""" , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam _lowerCAmelCase =None builder_instance.download_and_prepare() _lowerCAmelCase =builder_instance.as_dataset() assert ds @pytest.mark.integration def _lowerCamelCase(__UpperCamelCase ) -> Union[str, Any]: _lowerCAmelCase =dataset_module_factory("""wikipedia""" , cache_dir=__UpperCamelCase ) _lowerCAmelCase =import_main_class(dataset_module.module_path , dataset=__UpperCamelCase ) _lowerCAmelCase =builder_cls( cache_dir=__UpperCamelCase , config_name="""20220301.frr""" , hash=dataset_module.hash , ) _lowerCAmelCase =builder_instance.as_streaming_dataset() assert ds assert isinstance(__UpperCamelCase , __UpperCamelCase ) assert "train" in ds assert isinstance(ds["""train"""] , __UpperCamelCase ) assert next(iter(ds["""train"""] ) )
350
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowerCAmelCase ( self ) -> Union[str, Any]: _lowerCAmelCase =1 _lowerCAmelCase =3 _lowerCAmelCase =(32, 32) _lowerCAmelCase =floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__UpperCAmelCase ) return image @property def _lowerCAmelCase ( self ) -> Tuple: torch.manual_seed(0 ) _lowerCAmelCase =UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=__UpperCAmelCase , only_cross_attention=(True, True, False) , num_class_embeds=1_00 , ) return model @property def _lowerCAmelCase ( self ) -> Tuple: torch.manual_seed(0 ) _lowerCAmelCase =AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def _lowerCAmelCase ( self ) -> Optional[Any]: torch.manual_seed(0 ) _lowerCAmelCase =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) return CLIPTextModel(__UpperCAmelCase ) def _lowerCAmelCase ( self ) -> int: _lowerCAmelCase ="""cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase =self.dummy_cond_unet_upscale _lowerCAmelCase =DDPMScheduler() _lowerCAmelCase =DDIMScheduler(prediction_type="""v_prediction""" ) _lowerCAmelCase =self.dummy_vae _lowerCAmelCase =self.dummy_text_encoder _lowerCAmelCase =CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase =self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase =Image.fromarray(np.uinta(__UpperCAmelCase ) ).convert("""RGB""" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _lowerCAmelCase =StableDiffusionUpscalePipeline( unet=__UpperCAmelCase , low_res_scheduler=__UpperCAmelCase , scheduler=__UpperCAmelCase , vae=__UpperCAmelCase , text_encoder=__UpperCAmelCase , tokenizer=__UpperCAmelCase , max_noise_level=3_50 , ) _lowerCAmelCase =sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) _lowerCAmelCase ="""A painting of a squirrel eating a burger""" _lowerCAmelCase =torch.Generator(device=__UpperCAmelCase ).manual_seed(0 ) _lowerCAmelCase =sd_pipe( [prompt] , image=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase =output.images _lowerCAmelCase =torch.Generator(device=__UpperCAmelCase ).manual_seed(0 ) _lowerCAmelCase =sd_pipe( [prompt] , image=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , return_dict=__UpperCAmelCase , )[0] _lowerCAmelCase =image[0, -3:, -3:, -1] _lowerCAmelCase =image_from_tuple[0, -3:, -3:, -1] _lowerCAmelCase =low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) _lowerCAmelCase =np.array([0.3_1_1_3, 0.3_9_1_0, 0.4_2_7_2, 0.4_8_5_9, 0.5_0_6_1, 0.4_6_5_2, 0.5_3_6_2, 0.5_7_1_5, 0.5_6_6_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase ="""cpu""" # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase =self.dummy_cond_unet_upscale _lowerCAmelCase =DDPMScheduler() _lowerCAmelCase =DDIMScheduler(prediction_type="""v_prediction""" ) _lowerCAmelCase =self.dummy_vae _lowerCAmelCase =self.dummy_text_encoder _lowerCAmelCase =CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase =self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase =Image.fromarray(np.uinta(__UpperCAmelCase ) ).convert("""RGB""" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _lowerCAmelCase =StableDiffusionUpscalePipeline( unet=__UpperCAmelCase , low_res_scheduler=__UpperCAmelCase , scheduler=__UpperCAmelCase , vae=__UpperCAmelCase , text_encoder=__UpperCAmelCase , tokenizer=__UpperCAmelCase , max_noise_level=3_50 , ) _lowerCAmelCase =sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) _lowerCAmelCase ="""A painting of a squirrel eating a burger""" _lowerCAmelCase =sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase =output.images assert image.shape[0] == 2 _lowerCAmelCase =torch.Generator(device=__UpperCAmelCase ).manual_seed(0 ) _lowerCAmelCase =sd_pipe( [prompt] , image=__UpperCAmelCase , generator=__UpperCAmelCase , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _lowerCAmelCase =output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase =self.dummy_cond_unet_upscale _lowerCAmelCase =DDPMScheduler() _lowerCAmelCase =DDIMScheduler(prediction_type="""v_prediction""" ) _lowerCAmelCase =self.dummy_vae _lowerCAmelCase =self.dummy_text_encoder _lowerCAmelCase =CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase =self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase =Image.fromarray(np.uinta(__UpperCAmelCase ) ).convert("""RGB""" ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 _lowerCAmelCase =unet.half() _lowerCAmelCase =text_encoder.half() # make sure here that pndm scheduler skips prk _lowerCAmelCase =StableDiffusionUpscalePipeline( unet=__UpperCAmelCase , low_res_scheduler=__UpperCAmelCase , scheduler=__UpperCAmelCase , vae=__UpperCAmelCase , text_encoder=__UpperCAmelCase , tokenizer=__UpperCAmelCase , max_noise_level=3_50 , ) _lowerCAmelCase =sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) _lowerCAmelCase ="""A painting of a squirrel eating a burger""" _lowerCAmelCase =torch.manual_seed(0 ) _lowerCAmelCase =sd_pipe( [prompt] , image=__UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=2 , output_type="""np""" , ).images _lowerCAmelCase =low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCAmelCase ( self ) -> Optional[Any]: _lowerCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _lowerCAmelCase =load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale""" """/upsampled_cat.npy""" ) _lowerCAmelCase ="""stabilityai/stable-diffusion-x4-upscaler""" _lowerCAmelCase =StableDiffusionUpscalePipeline.from_pretrained(__UpperCAmelCase ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) pipe.enable_attention_slicing() _lowerCAmelCase ="""a cat sitting on a park bench""" _lowerCAmelCase =torch.manual_seed(0 ) _lowerCAmelCase =pipe( prompt=__UpperCAmelCase , image=__UpperCAmelCase , generator=__UpperCAmelCase , output_type="""np""" , ) _lowerCAmelCase =output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 1e-3 def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _lowerCAmelCase =load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale""" """/upsampled_cat_fp16.npy""" ) _lowerCAmelCase ="""stabilityai/stable-diffusion-x4-upscaler""" _lowerCAmelCase =StableDiffusionUpscalePipeline.from_pretrained( __UpperCAmelCase , torch_dtype=torch.floataa , ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) pipe.enable_attention_slicing() _lowerCAmelCase ="""a cat sitting on a park bench""" _lowerCAmelCase =torch.manual_seed(0 ) _lowerCAmelCase =pipe( prompt=__UpperCAmelCase , image=__UpperCAmelCase , generator=__UpperCAmelCase , output_type="""np""" , ) _lowerCAmelCase =output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 5e-1 def _lowerCAmelCase ( self ) -> Optional[Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _lowerCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _lowerCAmelCase ="""stabilityai/stable-diffusion-x4-upscaler""" _lowerCAmelCase =StableDiffusionUpscalePipeline.from_pretrained( __UpperCAmelCase , torch_dtype=torch.floataa , ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _lowerCAmelCase ="""a cat sitting on a park bench""" _lowerCAmelCase =torch.manual_seed(0 ) _lowerCAmelCase =pipe( prompt=__UpperCAmelCase , image=__UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=5 , output_type="""np""" , ) _lowerCAmelCase =torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
341
0
'''simple docstring''' class UpperCAmelCase_ : """simple docstring""" def __init__( self , SCREAMING_SNAKE_CASE_ ) -> Dict: # we need a list not a string, so do something to change the type __lowerCamelCase : Tuple = arr.split(',' ) def lowercase_ ( self ) -> Dict: __lowerCamelCase : List[Any] = [int(self.array[0] )] * len(self.array ) __lowerCamelCase : int = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): __lowerCamelCase : List[str] = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) __lowerCamelCase : str = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": A__ : List[str] = input("""please input some numbers:""") A__ : Dict = SubArray(whole_array) A__ : List[Any] = array.solve_sub_array() print(("""the results is:""", re))
185
import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, AutoModelForMaskedLM, AutoTokenizer, DataCollatorForWholeWordMask, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process _snake_case = logging.getLogger(__name__) _snake_case = list(MODEL_FOR_MASKED_LM_MAPPING.keys()) _snake_case = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _snake_case : lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={ "help": ( "The model checkpoint for weights initialization.Don't set if you want to train a model from scratch." ) } , ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_lowercase )} , ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={ "help": ( "Override some existing default config settings when a model is trained from scratch. Example: " "n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index" ) } , ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) lowerCamelCase__: bool = field( default=_lowercase , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) lowerCamelCase__: str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCamelCase__: bool = field( default=_lowercase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) def _lowerCamelCase ( self: str ) -> Tuple: if self.config_overrides is not None and (self.config_name is not None or self.model_name_or_path is not None): raise ValueError( "--config_overrides can't be used in combination with --config_name or --model_name_or_path" ) @dataclass class _snake_case : lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "The name of the dataset to use (via the datasets library)."} ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCamelCase__: Optional[str] = field(default=_lowercase , metadata={"help": "The input training data file (a text file)."} ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "An optional input train ref data file for whole word masking in Chinese."} , ) lowerCamelCase__: Optional[str] = field( default=_lowercase , metadata={"help": "An optional input validation ref data file for whole word masking in Chinese."} , ) lowerCamelCase__: bool = field( default=_lowercase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) lowerCamelCase__: Optional[int] = field( default=5 , metadata={ "help": "The percentage of the train set used as validation set in case there's no validation split" } , ) lowerCamelCase__: Optional[int] = field( default=_lowercase , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated. Default to the max input length of the model." ) } , ) lowerCamelCase__: Optional[int] = field( default=_lowercase , metadata={"help": "The number of processes to use for the preprocessing."} , ) lowerCamelCase__: float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"} ) lowerCamelCase__: bool = field( default=_lowercase , metadata={ "help": ( "Whether to pad all samples to `max_seq_length`. " "If False, will pad the samples dynamically when batching to the maximum length in the batch." ) } , ) def _lowerCamelCase ( self: Any ) -> Tuple: if self.train_file is not None: __UpperCAmelCase : Optional[int] = self.train_file.split("." )[-1] assert extension in ["csv", "json", "txt"], "`train_file` should be a csv, a json or a txt file." if self.validation_file is not None: __UpperCAmelCase : str = self.validation_file.split("." )[-1] assert extension in ["csv", "json", "txt"], "`validation_file` should be a csv, a json or a txt file." def _UpperCamelCase ( snake_case__, snake_case__ ) -> Optional[int]: with open(snake_case__, "r", encoding="utf-8" ) as f: __UpperCAmelCase : List[str] = [json.loads(snake_case__ ) for line in f.read().splitlines() if (len(snake_case__ ) > 0 and not line.isspace())] assert len(snake_case__ ) == len(snake_case__ ) __UpperCAmelCase : Optional[int] = {c: dataset[c] for c in dataset.column_names} __UpperCAmelCase : Any = refs return Dataset.from_dict(snake_case__ ) def _UpperCamelCase ( ) -> str: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __UpperCAmelCase : Optional[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Any = parser.parse_args_into_dataclasses() # Detecting last checkpoint. __UpperCAmelCase : int = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __UpperCAmelCase : List[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", handlers=[logging.StreamHandler(sys.stdout )], ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("Training/evaluation parameters %s", snake_case__ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. __UpperCAmelCase : Optional[Any] = load_dataset(data_args.dataset_name, data_args.dataset_config_name ) if "validation" not in datasets.keys(): __UpperCAmelCase : Dict = load_dataset( data_args.dataset_name, data_args.dataset_config_name, split=f'''train[:{data_args.validation_split_percentage}%]''', ) __UpperCAmelCase : List[str] = load_dataset( data_args.dataset_name, data_args.dataset_config_name, split=f'''train[{data_args.validation_split_percentage}%:]''', ) else: __UpperCAmelCase : List[Any] = {} if data_args.train_file is not None: __UpperCAmelCase : Optional[int] = data_args.train_file if data_args.validation_file is not None: __UpperCAmelCase : List[str] = data_args.validation_file __UpperCAmelCase : Tuple = data_args.train_file.split("." )[-1] if extension == "txt": __UpperCAmelCase : str = "text" __UpperCAmelCase : List[Any] = load_dataset(snake_case__, data_files=snake_case__ ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __UpperCAmelCase : Tuple = { "cache_dir": model_args.cache_dir, "revision": model_args.model_revision, "use_auth_token": True if model_args.use_auth_token else None, } if model_args.config_name: __UpperCAmelCase : Any = AutoConfig.from_pretrained(model_args.config_name, **snake_case__ ) elif model_args.model_name_or_path: __UpperCAmelCase : int = AutoConfig.from_pretrained(model_args.model_name_or_path, **snake_case__ ) else: __UpperCAmelCase : str = CONFIG_MAPPING[model_args.model_type]() logger.warning("You are instantiating a new config instance from scratch." ) if model_args.config_overrides is not None: logger.info(f'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(f'''New config: {config}''' ) __UpperCAmelCase : List[Any] = { "cache_dir": model_args.cache_dir, "use_fast": model_args.use_fast_tokenizer, "revision": model_args.model_revision, "use_auth_token": True if model_args.use_auth_token else None, } if model_args.tokenizer_name: __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained(model_args.tokenizer_name, **snake_case__ ) elif model_args.model_name_or_path: __UpperCAmelCase : Optional[int] = AutoTokenizer.from_pretrained(model_args.model_name_or_path, **snake_case__ ) else: raise ValueError( "You are instantiating a new tokenizer from scratch. This is not supported by this script." "You can do it from another script, save it, and load it from here, using --tokenizer_name." ) if model_args.model_name_or_path: __UpperCAmelCase : int = AutoModelForMaskedLM.from_pretrained( model_args.model_name_or_path, from_tf=bool(".ckpt" in model_args.model_name_or_path ), config=snake_case__, cache_dir=model_args.cache_dir, revision=model_args.model_revision, use_auth_token=True if model_args.use_auth_token else None, ) else: logger.info("Training new model from scratch" ) __UpperCAmelCase : Any = AutoModelForMaskedLM.from_config(snake_case__ ) model.resize_token_embeddings(len(snake_case__ ) ) # Preprocessing the datasets. # First we tokenize all the texts. if training_args.do_train: __UpperCAmelCase : List[str] = datasets["train"].column_names else: __UpperCAmelCase : Union[str, Any] = datasets["validation"].column_names __UpperCAmelCase : Union[str, Any] = "text" if "text" in column_names else column_names[0] __UpperCAmelCase : Any = "max_length" if data_args.pad_to_max_length else False def tokenize_function(snake_case__ ): # Remove empty lines __UpperCAmelCase : Any = [line for line in examples["text"] if len(snake_case__ ) > 0 and not line.isspace()] return tokenizer(examples["text"], padding=snake_case__, truncation=snake_case__, max_length=data_args.max_seq_length ) __UpperCAmelCase : List[str] = datasets.map( snake_case__, batched=snake_case__, num_proc=data_args.preprocessing_num_workers, remove_columns=[text_column_name], load_from_cache_file=not data_args.overwrite_cache, ) # Add the chinese references if provided if data_args.train_ref_file is not None: __UpperCAmelCase : str = add_chinese_references(tokenized_datasets["train"], data_args.train_ref_file ) if data_args.validation_ref_file is not None: __UpperCAmelCase : List[str] = add_chinese_references( tokenized_datasets["validation"], data_args.validation_ref_file ) # If we have ref files, need to avoid it removed by trainer __UpperCAmelCase : List[str] = data_args.train_ref_file or data_args.validation_ref_file if has_ref: __UpperCAmelCase : Tuple = False # Data collator # This one will take care of randomly masking the tokens. __UpperCAmelCase : Optional[Any] = DataCollatorForWholeWordMask(tokenizer=snake_case__, mlm_probability=data_args.mlm_probability ) # Initialize our Trainer __UpperCAmelCase : str = Trainer( model=snake_case__, args=snake_case__, train_dataset=tokenized_datasets["train"] if training_args.do_train else None, eval_dataset=tokenized_datasets["validation"] if training_args.do_eval else None, tokenizer=snake_case__, data_collator=snake_case__, ) # Training if training_args.do_train: if last_checkpoint is not None: __UpperCAmelCase : int = last_checkpoint elif model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ): __UpperCAmelCase : Any = model_args.model_name_or_path else: __UpperCAmelCase : Tuple = None __UpperCAmelCase : str = trainer.train(resume_from_checkpoint=snake_case__ ) trainer.save_model() # Saves the tokenizer too for easy upload __UpperCAmelCase : str = os.path.join(training_args.output_dir, "train_results.txt" ) if trainer.is_world_process_zero(): with open(snake_case__, "w" ) as writer: logger.info("***** Train results *****" ) for key, value in sorted(train_result.metrics.items() ): logger.info(f''' {key} = {value}''' ) writer.write(f'''{key} = {value}\n''' ) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir, "trainer_state.json" ) ) # Evaluation __UpperCAmelCase : Dict = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) __UpperCAmelCase : List[Any] = trainer.evaluate() __UpperCAmelCase : int = math.exp(eval_output["eval_loss"] ) __UpperCAmelCase : Union[str, Any] = perplexity __UpperCAmelCase : List[Any] = os.path.join(training_args.output_dir, "eval_results_mlm_wwm.txt" ) if trainer.is_world_process_zero(): with open(snake_case__, "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in sorted(results.items() ): logger.info(f''' {key} = {value}''' ) writer.write(f'''{key} = {value}\n''' ) return results def _UpperCamelCase ( snake_case__ ) -> Dict: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
157
0
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging lowerCamelCase_ = logging.get_logger(__name__) logging.set_verbosity_info() def lowerCamelCase ( a_ , a_ ) -> int: if "xprophetnet" in prophetnet_checkpoint_path: lowerCAmelCase_ = XLMProphetNetForConditionalGenerationOld.from_pretrained(a_ ) lowerCAmelCase_ , lowerCAmelCase_ = XLMProphetNetForConditionalGeneration.from_pretrained( a_ , output_loading_info=a_ ) else: lowerCAmelCase_ = ProphetNetForConditionalGenerationOld.from_pretrained(a_ ) lowerCAmelCase_ , lowerCAmelCase_ = ProphetNetForConditionalGeneration.from_pretrained( a_ , output_loading_info=a_ ) lowerCAmelCase_ = ['key_proj', 'value_proj', 'query_proj'] lowerCAmelCase_ = { 'self_attn': 'ngram_self_attn', 'cross_attn': 'encoder_attn', 'cross_attn_layer_norm': 'encoder_attn_layer_norm', 'feed_forward_layer_norm': 'final_layer_norm', 'feed_forward': '', 'intermediate': 'fc1', 'output': 'fc2', 'key_proj': 'k_proj', 'query_proj': 'q_proj', 'value_proj': 'v_proj', 'word_embeddings': 'embed_tokens', 'embeddings_layer_norm': 'emb_layer_norm', 'relative_pos_embeddings': 'relative_linear', 'ngram_embeddings': 'ngram_input_embed', 'position_embeddings': 'embed_positions', } for key in loading_info["missing_keys"]: lowerCAmelCase_ = key.split('.' ) if attributes[0] == "lm_head": lowerCAmelCase_ = prophet lowerCAmelCase_ = prophet_old else: lowerCAmelCase_ = prophet.prophetnet lowerCAmelCase_ = prophet_old.model lowerCAmelCase_ = False for attribute in attributes: if attribute in mapping: lowerCAmelCase_ = mapping[attribute] if not hasattr(a_ , a_ ) and len(a_ ) > 0: lowerCAmelCase_ = attribute elif hasattr(a_ , a_ ): lowerCAmelCase_ = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" lowerCAmelCase_ = old_model.weight logger.info(F'''{attribute} is initialized.''' ) lowerCAmelCase_ = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" lowerCAmelCase_ = old_model.bias logger.info(F'''{attribute} is initialized''' ) lowerCAmelCase_ = True break elif attribute in special_keys and hasattr(a_ , 'in_proj_weight' ): lowerCAmelCase_ = old_model.in_proj_weight.shape[0] // 3 lowerCAmelCase_ = getattr(a_ , a_ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": lowerCAmelCase_ = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) lowerCAmelCase_ = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": lowerCAmelCase_ = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) lowerCAmelCase_ = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": lowerCAmelCase_ = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) lowerCAmelCase_ = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) lowerCAmelCase_ = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." lowerCAmelCase_ = nn.Parameter(old_model.embed_positions.weight[:512, :] ) lowerCAmelCase_ = True break if attribute.isdigit(): lowerCAmelCase_ = model[int(a_ )] lowerCAmelCase_ = old_model[int(a_ )] else: lowerCAmelCase_ = getattr(a_ , a_ ) if old_attribute == "": lowerCAmelCase_ = old_model else: if not hasattr(a_ , a_ ): raise ValueError(F'''{old_model} does not have {old_attribute}''' ) lowerCAmelCase_ = getattr(a_ , a_ ) if not is_key_init: raise ValueError(F'''{key} was not correctly initialized!''' ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) prophet.save_pretrained(a_ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowerCamelCase_ = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
14
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def lowerCamelCase ( a_ , a_ , a_ , a_ , a_ ) -> List[Any]: # load base model lowerCAmelCase_ = StableDiffusionPipeline.from_pretrained(a_ , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors lowerCAmelCase_ = load_file(a_ ) lowerCAmelCase_ = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: lowerCAmelCase_ = key.split('.' )[0].split(LORA_PREFIX_TEXT_ENCODER + '_' )[-1].split('_' ) lowerCAmelCase_ = pipeline.text_encoder else: lowerCAmelCase_ = key.split('.' )[0].split(LORA_PREFIX_UNET + '_' )[-1].split('_' ) lowerCAmelCase_ = pipeline.unet # find the target layer lowerCAmelCase_ = layer_infos.pop(0 ) while len(a_ ) > -1: try: lowerCAmelCase_ = curr_layer.__getattr__(a_ ) if len(a_ ) > 0: lowerCAmelCase_ = layer_infos.pop(0 ) elif len(a_ ) == 0: break except Exception: if len(a_ ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: lowerCAmelCase_ = layer_infos.pop(0 ) lowerCAmelCase_ = [] if "lora_down" in key: pair_keys.append(key.replace('lora_down' , 'lora_up' ) ) pair_keys.append(a_ ) else: pair_keys.append(a_ ) pair_keys.append(key.replace('lora_up' , 'lora_down' ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: lowerCAmelCase_ = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) lowerCAmelCase_ = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(a_ , a_ ).unsqueeze(2 ).unsqueeze(3 ) else: lowerCAmelCase_ = state_dict[pair_keys[0]].to(torch.floataa ) lowerCAmelCase_ = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(a_ , a_ ) # update visited list for item in pair_keys: visited.append(a_ ) return pipeline if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( """--base_model_path""", default=None, type=str, required=True, help="""Path to the base model in diffusers format.""" ) parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--lora_prefix_unet""", default="""lora_unet""", type=str, help="""The prefix of UNet weight in safetensors""" ) parser.add_argument( """--lora_prefix_text_encoder""", default="""lora_te""", type=str, help="""The prefix of text encoder weight in safetensors""", ) parser.add_argument("""--alpha""", default=0.75, type=float, help="""The merging ratio in W = W0 + alpha * deltaW""") parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""" ) parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") lowerCamelCase_ = parser.parse_args() lowerCamelCase_ = args.base_model_path lowerCamelCase_ = args.checkpoint_path lowerCamelCase_ = args.dump_path lowerCamelCase_ = args.lora_prefix_unet lowerCamelCase_ = args.lora_prefix_text_encoder lowerCamelCase_ = args.alpha lowerCamelCase_ = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) lowerCamelCase_ = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
14
1
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) lowerCAmelCase_ = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', F'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', F'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', F'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', F'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_v_proj.weight''', F'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', F'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', F'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', F'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_v_proj.weight''', F'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', F'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', F'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', F'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', F'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', F'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_v_proj.bias''', F'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', F'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', F'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', F'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_v_proj.bias''', F'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', F'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" snake_case_ : Tuple = state_dict.pop(_UpperCamelCase ) snake_case_ : str = val def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" snake_case_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Dict = key.replace('''backbone.0.body''' , '''backbone.conv_encoder.model''' ) snake_case_ : str = value else: snake_case_ : Union[str, Any] = value return new_state_dict def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase=False ) -> Union[str, Any]: """simple docstring""" snake_case_ : Union[str, Any] = '''''' if is_panoptic: snake_case_ : Optional[int] = '''conditional_detr.''' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Tuple = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : List[Any] = in_proj_weight[:256, :] snake_case_ : List[str] = in_proj_bias[:256] snake_case_ : Tuple = in_proj_weight[256:512, :] snake_case_ : Union[str, Any] = in_proj_bias[256:512] snake_case_ : Optional[Any] = in_proj_weight[-256:, :] snake_case_ : List[Any] = in_proj_bias[-256:] def lowerCamelCase_ ( ) -> Optional[Any]: """simple docstring""" snake_case_ : List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case_ : Union[str, Any] = Image.open(requests.get(_UpperCamelCase , stream=_UpperCamelCase ).raw ) return im @torch.no_grad() def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Optional[int]: """simple docstring""" snake_case_ : Dict = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : List[Any] = '''resnet101''' if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : str = '''panoptic''' in model_name if is_panoptic: snake_case_ : str = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[Any] = '''huggingface/label-files''' snake_case_ : Union[str, Any] = '''coco-detection-id2label.json''' snake_case_ : Union[str, Any] = json.load(open(hf_hub_download(_UpperCamelCase , _UpperCamelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case_ : Optional[int] = {int(_UpperCamelCase ): v for k, v in idalabel.items()} snake_case_ : Tuple = idalabel snake_case_ : List[str] = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : str = '''coco_panoptic''' if is_panoptic else '''coco_detection''' snake_case_ : Optional[Any] = ConditionalDetrImageProcessor(format=_UpperCamelCase ) # prepare image snake_case_ : Dict = prepare_img() snake_case_ : Dict = image_processor(images=_UpperCamelCase , return_tensors='''pt''' ) snake_case_ : Union[str, Any] = encoding['''pixel_values'''] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Any = torch.hub.load('''DeppMeng/ConditionalDETR''' , _UpperCamelCase , pretrained=_UpperCamelCase ).eval() snake_case_ : Optional[Any] = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : List[str] = '''conditional_detr.''' + src rename_key(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) snake_case_ : int = rename_backbone_keys(_UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(_UpperCamelCase , is_panoptic=_UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : str = '''conditional_detr.model.''' if is_panoptic else '''model.''' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('''conditional_detr''' ) and not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ) ): snake_case_ : int = state_dict.pop(_UpperCamelCase ) snake_case_ : Dict = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Optional[Any] = state_dict.pop(_UpperCamelCase ) snake_case_ : List[str] = val elif key.startswith('''bbox_attention''' ) or key.startswith('''mask_head''' ): continue else: snake_case_ : List[Any] = state_dict.pop(_UpperCamelCase ) snake_case_ : Tuple = val else: if not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ): snake_case_ : Any = state_dict.pop(_UpperCamelCase ) snake_case_ : Union[str, Any] = val # finally, create HuggingFace model and load state dict snake_case_ : int = ConditionalDetrForSegmentation(_UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(_UpperCamelCase ) model.load_state_dict(_UpperCamelCase ) model.eval() model.push_to_hub(repo_id=_UpperCamelCase , organization='''DepuMeng''' , commit_message='''Add model''' ) # verify our conversion snake_case_ : Any = conditional_detr(_UpperCamelCase ) snake_case_ : Union[str, Any] = model(_UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs['''pred_logits'''] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs['''pred_boxes'''] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['''pred_masks'''] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) model.save_pretrained(_UpperCamelCase ) image_processor.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) lowerCAmelCase_ = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
279
import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets lowerCAmelCase_ = datasets.logging.get_logger(__name__) lowerCAmelCase_ = '''\ @inproceedings{bleurt, title={BLEURT: Learning Robust Metrics for Text Generation}, author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh}, booktitle={ACL}, year={2020}, url={https://arxiv.org/abs/2004.04696} } ''' lowerCAmelCase_ = '''\ BLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018) and then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune it for your specific application (the latter is expected to perform better). See the project\'s README at https://github.com/google-research/bleurt#readme for more information. ''' lowerCAmelCase_ = ''' BLEURT score. Args: `predictions` (list of str): prediction/candidate sentences `references` (list of str): reference sentences `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None. Returns: \'scores\': List of scores. Examples: >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> bleurt = datasets.load_metric("bleurt") >>> results = bleurt.compute(predictions=predictions, references=references) >>> print([round(v, 2) for v in results["scores"]]) [1.03, 1.04] ''' lowerCAmelCase_ = { '''bleurt-tiny-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip''', '''bleurt-tiny-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip''', '''bleurt-base-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip''', '''bleurt-base-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip''', '''bleurt-large-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip''', '''bleurt-large-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip''', '''BLEURT-20-D3''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip''', '''BLEURT-20-D6''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip''', '''BLEURT-20-D12''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip''', '''BLEURT-20''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip''', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/google-research/bleurt''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/bleurt'''] , reference_urls=['''https://github.com/google-research/bleurt''', '''https://arxiv.org/abs/2004.04696'''] , ) def lowerCamelCase (self , __magic_name__ ) -> List[Any]: '''simple docstring''' if self.config_name == "default": logger.warning( '''Using default BLEURT-Base checkpoint for sequence maximum length 128. ''' '''You can use a bigger model for better results with e.g.: datasets.load_metric(\'bleurt\', \'bleurt-large-512\').''' ) snake_case_ : Dict = '''bleurt-base-128''' if self.config_name.lower() in CHECKPOINT_URLS: snake_case_ : Optional[int] = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: snake_case_ : Union[str, Any] = self.config_name.upper() else: raise KeyError( F'''{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}''' ) # download the model checkpoint specified by self.config_name and set up the scorer snake_case_ : Any = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) snake_case_ : Dict = score.BleurtScorer(os.path.join(__magic_name__ , __magic_name__ ) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = self.scorer.score(references=__magic_name__ , candidates=__magic_name__ ) return {"scores": scores}
279
1
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : bool = False ) -> list[float]: if radian_mode: return [magnitude * cos(lowercase_ ), magnitude * sin(lowercase_ )] return [magnitude * cos(radians(lowercase_ ) ), magnitude * sin(radians(lowercase_ ) )] def lowerCAmelCase_( lowercase_ : NDArray[floataa] , lowercase_ : NDArray[floataa] , lowercase_ : float = 10**-1 ) -> bool: _lowerCamelCase = cross(lowercase_ , lowercase_ ) _lowerCamelCase = sum(lowercase_ ) return abs(lowercase_ ) < eps if __name__ == "__main__": # Test to check if it works __SCREAMING_SNAKE_CASE : Union[str, Any] = array( [ polar_force(718.4, 1_8_0 - 3_0), polar_force(879.54, 4_5), polar_force(1_0_0, -9_0), ] ) __SCREAMING_SNAKE_CASE : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg __SCREAMING_SNAKE_CASE : Optional[int] = array( [ polar_force(3_0 * 9.81, 1_5), polar_force(2_1_5, 1_8_0 - 4_5), polar_force(2_6_4, 9_0 - 3_0), ] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg __SCREAMING_SNAKE_CASE : str = array([[0, -2_0_0_0], [0, -1_2_0_0], [0, 1_5_6_0_0], [0, -1_2_4_0_0]]) __SCREAMING_SNAKE_CASE : str = array([[0, 0], [6, 0], [1_0, 0], [1_2, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
73
"""simple docstring""" import os from collections.abc import Iterator def lowerCAmelCase_( lowercase_ : str = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(lowercase_ ): _lowerCamelCase = [d for d in dir_names if d != '''scripts''' and d[0] not in '''._'''] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowercase_ )[1] in (".py", ".ipynb"): yield os.path.join(lowercase_ , lowercase_ ).lstrip('''./''' ) def lowerCAmelCase_( lowercase_ : Dict ) -> List[Any]: return F"""{i * " "}*""" if i else "\n##" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowercase_ ) or old_parts[i] != new_part) and new_part: print(F"""{md_prefix(lowercase_ )} {new_part.replace("_" , " " ).title()}""" ) return new_path def lowerCAmelCase_( lowercase_ : str = "." ) -> None: _lowerCamelCase = '''''' for filepath in sorted(good_file_paths(lowercase_ ) ): _lowerCamelCase , _lowerCamelCase = os.path.split(lowercase_ ) if filepath != old_path: _lowerCamelCase = print_path(lowercase_ , lowercase_ ) _lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 _lowerCamelCase = F"""{filepath}/{filename}""".replace(''' ''' , '''%20''' ) _lowerCamelCase = os.path.splitext(filename.replace('''_''' , ''' ''' ).title() )[0] print(F"""{md_prefix(lowercase_ )} [{filename}]({url})""" ) if __name__ == "__main__": print_directory_md('''.''')
73
1
from collections import defaultdict from math import ceil, sqrt def _UpperCamelCase ( lowercase__ = 1000000 , lowercase__ = 10 ): __SCREAMING_SNAKE_CASE : defaultdict = defaultdict(lowercase__ ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: __SCREAMING_SNAKE_CASE : Tuple = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(lowercase__ , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(f"""{solution() = }""")
9
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class _lowercase ( A__ ): '''simple docstring''' def __init__( self :int , lowerCAmelCase__ :NestedDataStructureLike[PathLike] , lowerCAmelCase__ :Optional[NamedSplit] = None , lowerCAmelCase__ :Optional[Features] = None , lowerCAmelCase__ :str = None , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :Optional[int] = None , **lowerCAmelCase__ :Optional[int] , ) -> Tuple: super().__init__( lowerCAmelCase__ , split=lowerCAmelCase__ , features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ , streaming=lowerCAmelCase__ , num_proc=lowerCAmelCase__ , **lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE : List[str] = path_or_paths if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else {self.split: path_or_paths} __SCREAMING_SNAKE_CASE : int = Text( cache_dir=lowerCAmelCase__ , data_files=lowerCAmelCase__ , features=lowerCAmelCase__ , **lowerCAmelCase__ , ) def __magic_name__( self :Dict ) -> Tuple: # Build iterable dataset if self.streaming: __SCREAMING_SNAKE_CASE : int = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __SCREAMING_SNAKE_CASE : List[str] = None __SCREAMING_SNAKE_CASE : str = None __SCREAMING_SNAKE_CASE : Dict = None __SCREAMING_SNAKE_CASE : Tuple = None self.builder.download_and_prepare( download_config=lowerCAmelCase__ , download_mode=lowerCAmelCase__ , verification_mode=lowerCAmelCase__ , base_path=lowerCAmelCase__ , num_proc=self.num_proc , ) __SCREAMING_SNAKE_CASE : Optional[int] = self.builder.as_dataset( split=self.split , verification_mode=lowerCAmelCase__ , in_memory=self.keep_in_memory ) return dataset
9
1
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = 42 class A ( nn.Module ): '''simple docstring''' def __init__(self : Optional[int] , _UpperCAmelCase : Any=3 , _UpperCAmelCase : Any=3 , _UpperCAmelCase : Any=("DownEncoderBlock2D",) , _UpperCAmelCase : List[str]=(64,) , _UpperCAmelCase : Optional[Any]=2 , _UpperCAmelCase : List[Any]=32 , _UpperCAmelCase : Optional[Any]="silu" , _UpperCAmelCase : List[str]=True , ) -> Optional[int]: """simple docstring""" super().__init__() lowercase__ = layers_per_block lowercase__ = torch.nn.Convad( _UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) lowercase__ = None lowercase__ = nn.ModuleList([] ) # down lowercase__ = block_out_channels[0] for i, down_block_type in enumerate(_UpperCAmelCase ): lowercase__ = output_channel lowercase__ = block_out_channels[i] lowercase__ = i == len(_UpperCAmelCase ) - 1 lowercase__ = get_down_block( _UpperCAmelCase , num_layers=self.layers_per_block , in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=_UpperCAmelCase , resnet_groups=_UpperCAmelCase , attention_head_dim=_UpperCAmelCase , temb_channels=_UpperCAmelCase , ) self.down_blocks.append(_UpperCAmelCase ) # mid lowercase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=_UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=_UpperCAmelCase , temb_channels=_UpperCAmelCase , ) # out lowercase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=_UpperCAmelCase , eps=1E-6 ) lowercase__ = nn.SiLU() lowercase__ = 2 * out_channels if double_z else out_channels lowercase__ = nn.Convad(block_out_channels[-1] , _UpperCAmelCase , 3 , padding=1 ) lowercase__ = False def lowerCamelCase__ (self : Optional[int] , _UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" lowercase__ = x lowercase__ = self.conv_in(_UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(_UpperCAmelCase : List[Any] ): def custom_forward(*_UpperCAmelCase : List[Any] ): return module(*_UpperCAmelCase ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: lowercase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(_UpperCAmelCase ) , _UpperCAmelCase , use_reentrant=_UpperCAmelCase ) # middle lowercase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , _UpperCAmelCase , use_reentrant=_UpperCAmelCase ) else: for down_block in self.down_blocks: lowercase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(_UpperCAmelCase ) , _UpperCAmelCase ) # middle lowercase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , _UpperCAmelCase ) else: # down for down_block in self.down_blocks: lowercase__ = down_block(_UpperCAmelCase ) # middle lowercase__ = self.mid_block(_UpperCAmelCase ) # post-process lowercase__ = self.conv_norm_out(_UpperCAmelCase ) lowercase__ = self.conv_act(_UpperCAmelCase ) lowercase__ = self.conv_out(_UpperCAmelCase ) return sample class A ( nn.Module ): '''simple docstring''' def __init__(self : Optional[int] , _UpperCAmelCase : Optional[Any]=3 , _UpperCAmelCase : List[str]=3 , _UpperCAmelCase : Union[str, Any]=("UpDecoderBlock2D",) , _UpperCAmelCase : Tuple=(64,) , _UpperCAmelCase : Union[str, Any]=2 , _UpperCAmelCase : Dict=32 , _UpperCAmelCase : Union[str, Any]="silu" , _UpperCAmelCase : Optional[Any]="group" , ) -> Any: """simple docstring""" super().__init__() lowercase__ = layers_per_block lowercase__ = nn.Convad( _UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) lowercase__ = None lowercase__ = nn.ModuleList([] ) lowercase__ = in_channels if norm_type == """spatial""" else None # mid lowercase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=_UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=_UpperCAmelCase , temb_channels=_UpperCAmelCase , ) # up lowercase__ = list(reversed(_UpperCAmelCase ) ) lowercase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(_UpperCAmelCase ): lowercase__ = output_channel lowercase__ = reversed_block_out_channels[i] lowercase__ = i == len(_UpperCAmelCase ) - 1 lowercase__ = get_up_block( _UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , prev_output_channel=_UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=_UpperCAmelCase , resnet_groups=_UpperCAmelCase , attention_head_dim=_UpperCAmelCase , temb_channels=_UpperCAmelCase , resnet_time_scale_shift=_UpperCAmelCase , ) self.up_blocks.append(_UpperCAmelCase ) lowercase__ = output_channel # out if norm_type == "spatial": lowercase__ = SpatialNorm(block_out_channels[0] , _UpperCAmelCase ) else: lowercase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=_UpperCAmelCase , eps=1E-6 ) lowercase__ = nn.SiLU() lowercase__ = nn.Convad(block_out_channels[0] , _UpperCAmelCase , 3 , padding=1 ) lowercase__ = False def lowerCamelCase__ (self : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" lowercase__ = z lowercase__ = self.conv_in(_UpperCAmelCase ) lowercase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(_UpperCAmelCase : Union[str, Any] ): def custom_forward(*_UpperCAmelCase : Dict ): return module(*_UpperCAmelCase ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle lowercase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , _UpperCAmelCase , _UpperCAmelCase , use_reentrant=_UpperCAmelCase ) lowercase__ = sample.to(_UpperCAmelCase ) # up for up_block in self.up_blocks: lowercase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(_UpperCAmelCase ) , _UpperCAmelCase , _UpperCAmelCase , use_reentrant=_UpperCAmelCase ) else: # middle lowercase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , _UpperCAmelCase , _UpperCAmelCase ) lowercase__ = sample.to(_UpperCAmelCase ) # up for up_block in self.up_blocks: lowercase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(_UpperCAmelCase ) , _UpperCAmelCase , _UpperCAmelCase ) else: # middle lowercase__ = self.mid_block(_UpperCAmelCase , _UpperCAmelCase ) lowercase__ = sample.to(_UpperCAmelCase ) # up for up_block in self.up_blocks: lowercase__ = up_block(_UpperCAmelCase , _UpperCAmelCase ) # post-process if latent_embeds is None: lowercase__ = self.conv_norm_out(_UpperCAmelCase ) else: lowercase__ = self.conv_norm_out(_UpperCAmelCase , _UpperCAmelCase ) lowercase__ = self.conv_act(_UpperCAmelCase ) lowercase__ = self.conv_out(_UpperCAmelCase ) return sample class A ( nn.Module ): '''simple docstring''' def __init__(self : List[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any]=None , _UpperCAmelCase : Dict="random" , _UpperCAmelCase : str=False , _UpperCAmelCase : int=True ) -> Optional[Any]: """simple docstring""" super().__init__() lowercase__ = n_e lowercase__ = vq_embed_dim lowercase__ = beta lowercase__ = legacy lowercase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) lowercase__ = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) lowercase__ = self.used.shape[0] lowercase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": lowercase__ = self.re_embed lowercase__ = self.re_embed + 1 print( f'''Remapping {self.n_e} indices to {self.re_embed} indices. ''' f'''Using {self.unknown_index} for unknown indices.''' ) else: lowercase__ = n_e lowercase__ = sane_index_shape def lowerCamelCase__ (self : Any , _UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" lowercase__ = inds.shape assert len(_UpperCAmelCase ) > 1 lowercase__ = inds.reshape(ishape[0] , -1 ) lowercase__ = self.used.to(_UpperCAmelCase ) lowercase__ = (inds[:, :, None] == used[None, None, ...]).long() lowercase__ = match.argmax(-1 ) lowercase__ = match.sum(2 ) < 1 if self.unknown_index == "random": lowercase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: lowercase__ = self.unknown_index return new.reshape(_UpperCAmelCase ) def lowerCamelCase__ (self : Union[str, Any] , _UpperCAmelCase : Union[str, Any] ) -> Any: """simple docstring""" lowercase__ = inds.shape assert len(_UpperCAmelCase ) > 1 lowercase__ = inds.reshape(ishape[0] , -1 ) lowercase__ = self.used.to(_UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token lowercase__ = 0 # simply set to zero lowercase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , _UpperCAmelCase ) return back.reshape(_UpperCAmelCase ) def lowerCamelCase__ (self : int , _UpperCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" lowercase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() lowercase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z lowercase__ = torch.argmin(torch.cdist(_UpperCAmelCase , self.embedding.weight ) , dim=1 ) lowercase__ = self.embedding(_UpperCAmelCase ).view(z.shape ) lowercase__ = None lowercase__ = None # compute loss for embedding if not self.legacy: lowercase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: lowercase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients lowercase__ = z + (z_q - z).detach() # reshape back to match original input shape lowercase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: lowercase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis lowercase__ = self.remap_to_used(_UpperCAmelCase ) lowercase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: lowercase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase__ (self : List[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any ) -> Any: """simple docstring""" if self.remap is not None: lowercase__ = indices.reshape(shape[0] , -1 ) # add batch axis lowercase__ = self.unmap_to_all(_UpperCAmelCase ) lowercase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors lowercase__ = self.embedding(_UpperCAmelCase ) if shape is not None: lowercase__ = z_q.view(_UpperCAmelCase ) # reshape back to match original input shape lowercase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase__ ): '''simple docstring''' def __init__(self : List[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int=False ) -> Union[str, Any]: """simple docstring""" lowercase__ = parameters lowercase__ , lowercase__ = torch.chunk(_UpperCAmelCase , 2 , dim=1 ) lowercase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) lowercase__ = deterministic lowercase__ = torch.exp(0.5 * self.logvar ) lowercase__ = torch.exp(self.logvar ) if self.deterministic: lowercase__ = lowercase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase__ (self : Any , _UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" lowercase__ = randn_tensor( self.mean.shape , generator=_UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) lowercase__ = self.mean + self.std * sample return x def lowerCamelCase__ (self : int , _UpperCAmelCase : Any=None ) -> List[str]: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase__ (self : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : str=[1, 2, 3] ) -> List[Any]: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) lowercase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=_UpperCAmelCase ) def lowerCamelCase__ (self : List[Any] ) -> List[str]: """simple docstring""" return self.mean
146
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class A ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase__ (self : Any ) -> Union[str, Any]: """simple docstring""" lowercase__ = TFXLMRobertaModel.from_pretrained("""jplu/tf-xlm-roberta-base""" ) lowercase__ = { """input_ids""": tf.convert_to_tensor([[0, 2646, 1_0269, 83, 9_9942, 2]] , dtype=tf.intaa ), # "My dog is cute" """attention_mask""": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } lowercase__ = model(_UpperCAmelCase )["""last_hidden_state"""] lowercase__ = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , _UpperCAmelCase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [ [ [0.0_681_762, 0.10_894_451, 0.06_772_504], [-0.06_423_668, 0.02_366_615, 0.04_329_344], [-0.06_057_295, 0.09_974_135, -0.00_070_584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
146
1
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position A_ : Dict = '2.13.1' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('3.7'): raise ImportWarning( 'To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( 'To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n' 'If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip A_ : Tuple = concatenate_datasets A_ : Union[str, Any] = DownloadConfig A_ : Any = DownloadManager A_ : Any = DownloadMode A_ : int = DownloadConfig A_ : int = DownloadMode A_ : Dict = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
333
import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ ( _a ): '''simple docstring''' a__ = (IPNDMScheduler,) a__ = (("num_inference_steps", 50),) def lowerCAmelCase_ (self , **lowercase__ ) -> Tuple: __UpperCAmelCase = {'''num_train_timesteps''': 1_000} config.update(**lowercase__ ) return config def lowerCAmelCase_ (self , lowercase__=0 , **lowercase__ ) -> Any: __UpperCAmelCase = dict(self.forward_default_kwargs ) __UpperCAmelCase = kwargs.pop('''num_inference_steps''' , lowercase__ ) __UpperCAmelCase = self.dummy_sample __UpperCAmelCase = 0.1 * sample __UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: __UpperCAmelCase = self.get_scheduler_config(**lowercase__ ) __UpperCAmelCase = scheduler_class(**lowercase__ ) scheduler.set_timesteps(lowercase__ ) # copy over dummy past residuals __UpperCAmelCase = dummy_past_residuals[:] if time_step is None: __UpperCAmelCase = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase__ ) __UpperCAmelCase = scheduler_class.from_pretrained(lowercase__ ) new_scheduler.set_timesteps(lowercase__ ) # copy over dummy past residuals __UpperCAmelCase = dummy_past_residuals[:] __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = new_scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = new_scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ (self ) -> List[str]: pass def lowerCAmelCase_ (self , lowercase__=0 , **lowercase__ ) -> Optional[int]: __UpperCAmelCase = dict(self.forward_default_kwargs ) __UpperCAmelCase = kwargs.pop('''num_inference_steps''' , lowercase__ ) __UpperCAmelCase = self.dummy_sample __UpperCAmelCase = 0.1 * sample __UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: __UpperCAmelCase = self.get_scheduler_config() __UpperCAmelCase = scheduler_class(**lowercase__ ) scheduler.set_timesteps(lowercase__ ) # copy over dummy past residuals (must be after setting timesteps) __UpperCAmelCase = dummy_past_residuals[:] if time_step is None: __UpperCAmelCase = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase__ ) __UpperCAmelCase = scheduler_class.from_pretrained(lowercase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase__ ) # copy over dummy past residual (must be after setting timesteps) __UpperCAmelCase = dummy_past_residuals[:] __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = new_scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = new_scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ (self , **lowercase__ ) -> List[Any]: __UpperCAmelCase = self.scheduler_classes[0] __UpperCAmelCase = self.get_scheduler_config(**lowercase__ ) __UpperCAmelCase = scheduler_class(**lowercase__ ) __UpperCAmelCase = 10 __UpperCAmelCase = self.dummy_model() __UpperCAmelCase = self.dummy_sample_deter scheduler.set_timesteps(lowercase__ ) for i, t in enumerate(scheduler.timesteps ): __UpperCAmelCase = model(lowercase__ , lowercase__ ) __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): __UpperCAmelCase = model(lowercase__ , lowercase__ ) __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ ).prev_sample return sample def lowerCAmelCase_ (self ) -> Optional[Any]: __UpperCAmelCase = dict(self.forward_default_kwargs ) __UpperCAmelCase = kwargs.pop('''num_inference_steps''' , lowercase__ ) for scheduler_class in self.scheduler_classes: __UpperCAmelCase = self.get_scheduler_config() __UpperCAmelCase = scheduler_class(**lowercase__ ) __UpperCAmelCase = self.dummy_sample __UpperCAmelCase = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase__ , '''set_timesteps''' ): scheduler.set_timesteps(lowercase__ ) elif num_inference_steps is not None and not hasattr(lowercase__ , '''set_timesteps''' ): __UpperCAmelCase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) __UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] __UpperCAmelCase = dummy_past_residuals[:] __UpperCAmelCase = scheduler.timesteps[5] __UpperCAmelCase = scheduler.timesteps[6] __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample __UpperCAmelCase = scheduler.step(lowercase__ , lowercase__ , lowercase__ , **lowercase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowerCAmelCase_ (self ) -> List[Any]: for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=lowercase__ , time_step=lowercase__ ) def lowerCAmelCase_ (self ) -> Union[str, Any]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase__ , time_step=lowercase__ ) def lowerCAmelCase_ (self ) -> str: __UpperCAmelCase = self.full_loop() __UpperCAmelCase = torch.mean(torch.abs(lowercase__ ) ) assert abs(result_mean.item() - 2_540_529 ) < 10
333
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class __a ( UpperCAmelCase ): _a : List[str] = ['image_processor'] _a : Optional[Any] = 'SamImageProcessor' def __init__( self , _SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = self.image_processor _UpperCAmelCase = -10 _UpperCAmelCase = self.image_processor.size['longest_edge'] def __call__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> BatchEncoding: """simple docstring""" _UpperCAmelCase = self.image_processor( _SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) # pop arguments that are not used in the foward but used nevertheless _UpperCAmelCase = encoding_image_processor['original_sizes'] if hasattr(_SCREAMING_SNAKE_CASE , 'numpy' ): # Checks if Torch or TF tensor _UpperCAmelCase = original_sizes.numpy() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = self._check_and_preprocess_points( input_points=_SCREAMING_SNAKE_CASE , input_labels=_SCREAMING_SNAKE_CASE , input_boxes=_SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = self._normalize_and_convert( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , input_points=_SCREAMING_SNAKE_CASE , input_labels=_SCREAMING_SNAKE_CASE , input_boxes=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , ) return encoding_image_processor def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE="pt" , ) -> Optional[int]: """simple docstring""" if input_points is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): _UpperCAmelCase = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , original_sizes[0] ) for point in input_points ] else: _UpperCAmelCase = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for point, original_size in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _UpperCAmelCase , _UpperCAmelCase = self._pad_points_and_labels(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.array(_SCREAMING_SNAKE_CASE ) if input_labels is not None: _UpperCAmelCase = np.array(_SCREAMING_SNAKE_CASE ) if input_boxes is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): _UpperCAmelCase = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , original_sizes[0] , is_bounding_box=_SCREAMING_SNAKE_CASE ) for box in input_boxes ] else: _UpperCAmelCase = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , is_bounding_box=_SCREAMING_SNAKE_CASE ) for box, original_size in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] _UpperCAmelCase = np.array(_SCREAMING_SNAKE_CASE ) if input_boxes is not None: if return_tensors == "pt": _UpperCAmelCase = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # boxes batch size of 1 by default _UpperCAmelCase = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _UpperCAmelCase = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # boxes batch size of 1 by default _UpperCAmelCase = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'input_boxes': input_boxes} ) if input_points is not None: if return_tensors == "pt": _UpperCAmelCase = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default _UpperCAmelCase = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _UpperCAmelCase = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default _UpperCAmelCase = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'input_points': input_points} ) if input_labels is not None: if return_tensors == "pt": _UpperCAmelCase = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default _UpperCAmelCase = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _UpperCAmelCase = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default _UpperCAmelCase = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'input_labels': input_labels} ) return encoding_image_processor def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = max([point.shape[0] for point in input_points] ) _UpperCAmelCase = [] for i, point in enumerate(_SCREAMING_SNAKE_CASE ): if point.shape[0] != expected_nb_points: _UpperCAmelCase = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _UpperCAmelCase = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = processed_input_points return input_points, input_labels def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> np.ndarray: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = original_size _UpperCAmelCase , _UpperCAmelCase = self.image_processor._get_preprocess_shape(_SCREAMING_SNAKE_CASE , longest_edge=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = deepcopy(_SCREAMING_SNAKE_CASE ).astype(_SCREAMING_SNAKE_CASE ) if is_bounding_box: _UpperCAmelCase = coords.reshape(-1 , 2 , 2 ) _UpperCAmelCase = coords[..., 0] * (new_w / old_w) _UpperCAmelCase = coords[..., 1] * (new_h / old_h) if is_bounding_box: _UpperCAmelCase = coords.reshape(-1 , 4 ) return coords def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) -> Dict: """simple docstring""" if input_points is not None: if hasattr(_SCREAMING_SNAKE_CASE , 'numpy' ): # Checks for TF or Torch tensor _UpperCAmelCase = input_points.numpy().tolist() if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_points[0] , _SCREAMING_SNAKE_CASE ): raise ValueError('Input points must be a list of list of floating points.' ) _UpperCAmelCase = [np.array(_SCREAMING_SNAKE_CASE ) for input_point in input_points] else: _UpperCAmelCase = None if input_labels is not None: if hasattr(_SCREAMING_SNAKE_CASE , 'numpy' ): _UpperCAmelCase = input_labels.numpy().tolist() if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_labels[0] , _SCREAMING_SNAKE_CASE ): raise ValueError('Input labels must be a list of list integers.' ) _UpperCAmelCase = [np.array(_SCREAMING_SNAKE_CASE ) for label in input_labels] else: _UpperCAmelCase = None if input_boxes is not None: if hasattr(_SCREAMING_SNAKE_CASE , 'numpy' ): _UpperCAmelCase = input_boxes.numpy().tolist() if ( not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_boxes[0] , _SCREAMING_SNAKE_CASE ) or not isinstance(input_boxes[0][0] , _SCREAMING_SNAKE_CASE ) ): raise ValueError('Input boxes must be a list of list of list of floating points.' ) _UpperCAmelCase = [np.array(_SCREAMING_SNAKE_CASE ).astype(np.floataa ) for box in input_boxes] else: _UpperCAmelCase = None return input_points, input_labels, input_boxes @property def UpperCAmelCase__ ( self ) -> Dict: """simple docstring""" _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(_SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase__ ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" return self.image_processor.post_process_masks(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
185
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def lowerCAmelCase__ ( a__: NDArray[floataa] , a__: NDArray[floataa] , a__: list[int] , a__: int , ) -> list[float]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = coefficient_matrix.shape _UpperCAmelCase , _UpperCAmelCase = constant_matrix.shape if rowsa != colsa: _UpperCAmelCase = F'''Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}''' raise ValueError(a__ ) if colsa != 1: _UpperCAmelCase = F'''Constant matrix must be nx1 but received {rowsa}x{colsa}''' raise ValueError(a__ ) if rowsa != rowsa: _UpperCAmelCase = ( 'Coefficient and constant matrices dimensions must be nxn and nx1 but ' F'''received {rowsa}x{colsa} and {rowsa}x{colsa}''' ) raise ValueError(a__ ) if len(a__ ) != rowsa: _UpperCAmelCase = ( 'Number of initial values must be equal to number of rows in coefficient ' F'''matrix but received {len(a__ )} and {rowsa}''' ) raise ValueError(a__ ) if iterations <= 0: raise ValueError('Iterations must be at least 1' ) _UpperCAmelCase = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) _UpperCAmelCase , _UpperCAmelCase = table.shape strictly_diagonally_dominant(a__ ) # Iterates the whole matrix for given number of times for _ in range(a__ ): _UpperCAmelCase = [] for row in range(a__ ): _UpperCAmelCase = 0 for col in range(a__ ): if col == row: _UpperCAmelCase = table[row][col] elif col == cols - 1: _UpperCAmelCase = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] _UpperCAmelCase = (temp + val) / denom new_val.append(a__ ) _UpperCAmelCase = new_val return [float(a__ ) for i in new_val] def lowerCAmelCase__ ( a__: NDArray[floataa] ) -> bool: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = table.shape _UpperCAmelCase = True for i in range(0 , a__ ): _UpperCAmelCase = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError('Coefficient matrix is not strictly diagonally dominant' ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
185
1
'''simple docstring''' def __magic_name__( lowerCamelCase, lowerCamelCase): def get_matched_characters(lowerCamelCase, lowerCamelCase) -> str: __lowerCAmelCase = [] __lowerCAmelCase = min(len(_stra), len(_stra)) // 2 for i, l in enumerate(_stra): __lowerCAmelCase = int(max(0, i - limit)) __lowerCAmelCase = int(min(i + limit + 1, len(_stra))) if l in _stra[left:right]: matched.append(lowerCamelCase) __lowerCAmelCase = F"""{_stra[0:_stra.index(lowerCamelCase)]} {_stra[_stra.index(lowerCamelCase) + 1:]}""" return "".join(lowerCamelCase) # matching characters __lowerCAmelCase = get_matched_characters(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = get_matched_characters(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = len(lowerCamelCase) # transposition __lowerCAmelCase = ( len([(ca, ca) for ca, ca in zip(lowerCamelCase, lowerCamelCase) if ca != ca]) // 2 ) if not match_count: __lowerCAmelCase = 0.0 else: __lowerCAmelCase = ( 1 / 3 * ( match_count / len(lowerCamelCase) + match_count / len(lowerCamelCase) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters __lowerCAmelCase = 0 for ca, ca in zip(stra[:4], stra[:4]): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("""hello""", """world"""))
174
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( __A , __A , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Union[str, Any] = StableDiffusionXLImgaImgPipeline __UpperCamelCase : str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} __UpperCamelCase : List[str] = PipelineTesterMixin.required_optional_params - {'latents'} __UpperCamelCase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __UpperCamelCase : Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS __UpperCamelCase : int = IMAGE_TO_IMAGE_IMAGE_PARAMS def _snake_case (self ): torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__lowercase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCAmelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCAmelCase = CLIPTextModel(__lowercase ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__lowercase ) __lowerCAmelCase = CLIPTextModelWithProjection(__lowercase ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__lowercase ) __lowerCAmelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def _snake_case (self , __lowercase , __lowercase=0 ): __lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase ) ).to(__lowercase ) __lowerCAmelCase = image / 2 + 0.5 if str(__lowercase ).startswith('''mps''' ): __lowerCAmelCase = torch.manual_seed(__lowercase ) else: __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = sd_pipe(**__lowercase ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _snake_case (self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def _snake_case (self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def _snake_case (self ): pass def _snake_case (self ): __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) __lowerCAmelCase = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) # forward without prompt embeds __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 3 * ['''this is a negative prompt'''] __lowerCAmelCase = negative_prompt __lowerCAmelCase = 3 * [inputs['''prompt''']] __lowerCAmelCase = sd_pipe(**__lowercase ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 3 * ['''this is a negative prompt'''] __lowerCAmelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = sd_pipe.encode_prompt(__lowercase , negative_prompt=__lowercase ) __lowerCAmelCase = sd_pipe( **__lowercase , prompt_embeds=__lowercase , negative_prompt_embeds=__lowercase , pooled_prompt_embeds=__lowercase , negative_pooled_prompt_embeds=__lowercase , ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case (self , __lowercase , __lowercase="cpu" , __lowercase=torch.floataa , __lowercase=0 ): __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = np.random.RandomState(__lowercase ).standard_normal((1, 4, 64, 64) ) __lowerCAmelCase = torch.from_numpy(__lowercase ).to(device=__lowercase , dtype=__lowercase ) __lowerCAmelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def _snake_case (self ): __lowerCAmelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs(__lowercase ) __lowerCAmelCase = pipe(**__lowercase ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCAmelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
174
1
'''simple docstring''' import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCamelCase ( unittest.TestCase ): """simple docstring""" @property def SCREAMING_SNAKE_CASE_ ( self : Optional[Any]): """simple docstring""" torch.manual_seed(0) a : Dict = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def SCREAMING_SNAKE_CASE_ ( self : Dict): """simple docstring""" a : str = self.dummy_uncond_unet a : List[Any] = ScoreSdeVeScheduler() a : Any = ScoreSdeVePipeline(unet=lowercase_ , scheduler=lowercase_) sde_ve.to(lowercase_) sde_ve.set_progress_bar_config(disable=lowercase_) a : Dict = torch.manual_seed(0) a : Optional[int] = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=lowercase_).images a : int = torch.manual_seed(0) a : Any = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=lowercase_ , return_dict=lowercase_)[ 0 ] a : str = image[0, -3:, -3:, -1] a : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) a : Optional[Any] = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class UpperCamelCase ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self : List[Any]): """simple docstring""" a : Optional[Any] = "google/ncsnpp-church-256" a : List[str] = UNetaDModel.from_pretrained(lowercase_) a : int = ScoreSdeVeScheduler.from_pretrained(lowercase_) a : List[str] = ScoreSdeVePipeline(unet=lowercase_ , scheduler=lowercase_) sde_ve.to(lowercase_) sde_ve.set_progress_bar_config(disable=lowercase_) a : Any = torch.manual_seed(0) a : List[Any] = sde_ve(num_inference_steps=1_0 , output_type='numpy' , generator=lowercase_).images a : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 2_5_6, 2_5_6, 3) a : str = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
351
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( snake_case : list[int | float] , snake_case : int , snake_case : int ) -> int | float: """simple docstring""" if len(snake_case ) == 0: raise ValueError('find_max() arg is an empty sequence' ) if ( left >= len(snake_case ) or left < -len(snake_case ) or right >= len(snake_case ) or right < -len(snake_case ) ): raise IndexError('list index out of range' ) if left == right: return nums[left] a : Union[str, Any] = (left + right) >> 1 # the middle a : List[str] = find_max(snake_case , snake_case , snake_case ) # find max in range[left, mid] a : Dict = find_max(snake_case , mid + 1 , snake_case ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
345
0
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->str: '''simple docstring''' if not isinstance(_lowercase , _lowercase ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) a : Tuple = precision a : str = ceil(precision / 14 ) a : List[Any] = 42_6880 * Decimal(1_0005 ).sqrt() a : Union[str, Any] = 1 a : Dict = 1359_1409 a : Optional[int] = Decimal(_lowercase ) for k in range(1 , _lowercase ): a : int = factorial(6 * k ) // (factorial(3 * k ) * factorial(_lowercase ) ** 3) linear_term += 5_4514_0134 exponential_term *= -26_2537_4126_4076_8000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": a : Optional[Any] = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
105
def SCREAMING_SNAKE_CASE__ ( ) -> list[list[int]]: return [list(range(1000 - i ,-1000 - i ,-1 ) ) for i in range(1000 )] lowerCamelCase : List[Any] = generate_large_matrix() lowerCamelCase : Optional[int] = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def SCREAMING_SNAKE_CASE__ ( lowercase ) -> None: assert all(row == sorted(lowercase ,reverse=lowercase ) for row in grid ) assert all(list(lowercase ) == sorted(lowercase ,reverse=lowercase ) for col in zip(*lowercase ) ) def SCREAMING_SNAKE_CASE__ ( lowercase ) -> int: snake_case : Tuple = 0 snake_case : List[Any] = len(lowercase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: snake_case : Tuple = (left + right) // 2 snake_case : Dict = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: snake_case : List[Any] = mid + 1 else: snake_case : str = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(lowercase ) def SCREAMING_SNAKE_CASE__ ( lowercase ) -> int: snake_case : Union[str, Any] = 0 snake_case : Dict = len(grid[0] ) for i in range(len(lowercase ) ): snake_case : Tuple = find_negative_index(grid[i][:bound] ) total += bound return (len(lowercase ) * len(grid[0] )) - total def SCREAMING_SNAKE_CASE__ ( lowercase ) -> int: return len([number for row in grid for number in row if number < 0] ) def SCREAMING_SNAKE_CASE__ ( lowercase ) -> int: snake_case : Dict = 0 for row in grid: for i, number in enumerate(lowercase ): if number < 0: total += len(lowercase ) - i break return total def SCREAMING_SNAKE_CASE__ ( ) -> None: from timeit import timeit print("""Running benchmarks""" ) snake_case : List[Any] = ( """from __main__ import count_negatives_binary_search, """ """count_negatives_brute_force, count_negatives_brute_force_with_break, grid""" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): snake_case : int = timeit(f"""{func}(grid=grid)""" ,setup=lowercase ,number=500 ) print(f"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
124
0
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def __lowerCamelCase ( __snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" A__ : str =FileLock(str(tmpdir / """foo.lock""" ) ) A__ : List[Any] =FileLock(str(tmpdir / """foo.lock""" ) ) A__ : Dict =0.01 with locka.acquire(): with pytest.raises(lowerCAmelCase__ ): A__ : Optional[int] =time.time() locka.acquire(lowerCAmelCase__ ) assert time.time() - _start > timeout def __lowerCamelCase ( __snake_case : Optional[int] ) -> Dict: """simple docstring""" A__ : int ="""a""" * 1_000 + """.lock""" A__ : List[Any] =FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(""".lock""" ) assert not locka._lock_file.endswith(lowerCAmelCase__ ) assert len(os.path.basename(locka._lock_file ) ) <= 255 A__ : Tuple =FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(lowerCAmelCase__ ): locka.acquire(0 )
360
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available __snake_case : Any = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[int] = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[Any] = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys __snake_case : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
136
0
import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class _snake_case : def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): if dst_width < 0 or dst_height < 0: raise ValueError('''Destination width/height should be > 0''' ) a :List[Any] = img a :Any = img.shape[1] a :str = img.shape[0] a :Optional[Any] = dst_width a :int = dst_height a :Optional[int] = self.src_w / self.dst_w a :Dict = self.src_h / self.dst_h a :Optional[Any] = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 255 ) def SCREAMING_SNAKE_CASE__ ( self ): for i in range(self.dst_h ): for j in range(self.dst_w ): a :Dict = self.img[self.get_y(_lowerCamelCase )][self.get_x(_lowerCamelCase )] def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return int(self.ratio_x * x ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return int(self.ratio_y * y ) if __name__ == "__main__": snake_case , snake_case : List[str] = 8_00, 6_00 snake_case : Dict = imread('''image_data/lena.jpg''', 1) snake_case : int = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F"""Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}""", n.output ) waitKey(0) destroyAllWindows()
94
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _snake_case ( unittest.TestCase ): '''simple docstring''' @property def A__ ( self: Optional[int] ) -> int: torch.manual_seed(0 ) UpperCAmelCase_ : Union[str, Any] = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") ,up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") ,) return model @property def A__ ( self: Tuple ) -> Optional[Any]: torch.manual_seed(0 ) UpperCAmelCase_ : List[str] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=3 ,) return model @property def A__ ( self: Tuple ) -> Any: torch.manual_seed(0 ) UpperCAmelCase_ : int = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(lowerCamelCase_ ) def A__ ( self: str ) -> Optional[Any]: UpperCAmelCase_ : str = self.dummy_uncond_unet UpperCAmelCase_ : List[Any] = DDIMScheduler() UpperCAmelCase_ : List[Any] = self.dummy_vq_model UpperCAmelCase_ : Optional[int] = LDMPipeline(unet=lowerCamelCase_ ,vqvae=lowerCamelCase_ ,scheduler=lowerCamelCase_ ) ldm.to(lowerCamelCase_ ) ldm.set_progress_bar_config(disable=lowerCamelCase_ ) UpperCAmelCase_ : Any = torch.manual_seed(0 ) UpperCAmelCase_ : int = ldm(generator=lowerCamelCase_ ,num_inference_steps=2 ,output_type="""numpy""" ).images UpperCAmelCase_ : List[str] = torch.manual_seed(0 ) UpperCAmelCase_ : Union[str, Any] = ldm(generator=lowerCamelCase_ ,num_inference_steps=2 ,output_type="""numpy""" ,return_dict=lowerCamelCase_ )[0] UpperCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] UpperCAmelCase_ : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : str = np.array([0.8_5_1_2, 0.8_1_8, 0.6_4_1_1, 0.6_8_0_8, 0.4_4_6_5, 0.5_6_1_8, 0.4_6, 0.6_2_3_1, 0.5_1_7_2] ) UpperCAmelCase_ : Tuple = 1e-2 if torch_device != """mps""" else 3e-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class _snake_case ( unittest.TestCase ): '''simple docstring''' def A__ ( self: Optional[int] ) -> Optional[Any]: UpperCAmelCase_ : List[str] = LDMPipeline.from_pretrained("""CompVis/ldm-celebahq-256""" ) ldm.to(lowerCamelCase_ ) ldm.set_progress_bar_config(disable=lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = torch.manual_seed(0 ) UpperCAmelCase_ : Optional[int] = ldm(generator=lowerCamelCase_ ,num_inference_steps=5 ,output_type="""numpy""" ).images UpperCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) UpperCAmelCase_ : int = np.array([0.4_3_9_9, 0.4_4_9_7_5, 0.4_6_8_2_5, 0.4_7_4, 0.4_3_5_9, 0.4_5_8_1, 0.4_5_0_9_5, 0.4_3_4_1, 0.4_4_4_7] ) UpperCAmelCase_ : Union[str, Any] = 1e-2 if torch_device != """mps""" else 3e-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
345
0
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowerCAmelCase : List[str] = """Run commands across TPU VMs for initial setup before running `accelerate launch`.""" def A_ ( _UpperCAmelCase=None ): if subparsers is not None: SCREAMING_SNAKE_CASE_: Optional[Any] = subparsers.add_parser("tpu-config" , description=_description ) else: SCREAMING_SNAKE_CASE_: Optional[int] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments SCREAMING_SNAKE_CASE_: Any = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_UpperCAmelCase , default=_UpperCAmelCase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_UpperCAmelCase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_UpperCAmelCase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) SCREAMING_SNAKE_CASE_: Tuple = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_UpperCAmelCase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Tuple = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: SCREAMING_SNAKE_CASE_: Any = defaults.command_file if not args.command and defaults.commands is not None: SCREAMING_SNAKE_CASE_: int = defaults.commands if not args.tpu_name: SCREAMING_SNAKE_CASE_: List[str] = defaults.tpu_name if not args.tpu_zone: SCREAMING_SNAKE_CASE_: Dict = defaults.tpu_zone if args.accelerate_version == "dev": SCREAMING_SNAKE_CASE_: Tuple = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": SCREAMING_SNAKE_CASE_: str = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = f"accelerate=={args.accelerate_version}" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: SCREAMING_SNAKE_CASE_: Dict = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate SCREAMING_SNAKE_CASE_: Dict = ["cd /usr/share"] if args.install_accelerate: new_cmd += [f"pip install {args.accelerate_version}"] new_cmd += args.command SCREAMING_SNAKE_CASE_: List[str] = "; ".join(_UpperCAmelCase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess SCREAMING_SNAKE_CASE_: int = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f"Running {' '.join(_UpperCAmelCase )}" ) return subprocess.run(_UpperCAmelCase ) print("Successfully setup pod." ) def A_ ( ): SCREAMING_SNAKE_CASE_: List[str] = tpu_command_parser() SCREAMING_SNAKE_CASE_: List[Any] = parser.parse_args() tpu_command_launcher(_UpperCAmelCase )
127
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer lowerCAmelCase : Any = logging.get_logger(__name__) lowerCAmelCase : Dict = {"""vocab_file""": """vocab.txt"""} lowerCAmelCase : List[str] = { """vocab_file""": { """YituTech/conv-bert-base""": """https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt""", """YituTech/conv-bert-medium-small""": ( """https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt""" ), """YituTech/conv-bert-small""": """https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt""", } } lowerCAmelCase : List[Any] = { """YituTech/conv-bert-base""": 512, """YituTech/conv-bert-medium-small""": 512, """YituTech/conv-bert-small""": 512, } lowerCAmelCase : Tuple = { """YituTech/conv-bert-base""": {"""do_lower_case""": True}, """YituTech/conv-bert-medium-small""": {"""do_lower_case""": True}, """YituTech/conv-bert-small""": {"""do_lower_case""": True}, } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : str = VOCAB_FILES_NAMES _UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : List[Any] = PRETRAINED_INIT_CONFIGURATION _UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Dict = ConvBertTokenizer def __init__( self : Optional[Any] , lowerCAmelCase__ : Optional[int]=None , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : str=True , lowerCAmelCase__ : Any="[UNK]" , lowerCAmelCase__ : Optional[Any]="[SEP]" , lowerCAmelCase__ : Any="[PAD]" , lowerCAmelCase__ : Dict="[CLS]" , lowerCAmelCase__ : Dict="[MASK]" , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : str=None , **lowerCAmelCase__ : Dict , ): super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , **lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get("lowercase" , lowerCAmelCase__) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase__) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase__) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE_: Optional[int] = getattr(lowerCAmelCase__ , normalizer_state.pop("type")) SCREAMING_SNAKE_CASE_: Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE_: List[str] = strip_accents SCREAMING_SNAKE_CASE_: Optional[Any] = tokenize_chinese_chars SCREAMING_SNAKE_CASE_: Optional[int] = normalizer_class(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = do_lower_case def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Union[str, Any]=None): SCREAMING_SNAKE_CASE_: List[str] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : List[int] , lowerCAmelCase__ : Optional[List[int]] = None): SCREAMING_SNAKE_CASE_: Optional[int] = [self.sep_token_id] SCREAMING_SNAKE_CASE_: int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[str] = None): SCREAMING_SNAKE_CASE_: Any = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__) return tuple(lowerCAmelCase__)
127
1
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder __lowercase = """base_with_context""" def lowercase ( A_ , A_ )-> List[Any]: '''simple docstring''' a : List[str] = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) a : List[str] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_SCREAMING_SNAKE_CASE ) for lyr_num, lyr in enumerate(model.encoders ): a : Optional[int] = weights[F'''layers_{lyr_num}'''] a : List[str] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) a : int = ly_weight["attention"] a : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) a : int = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) a : Any = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) a : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) a : int = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) a : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) a : int = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) a : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) a : List[Any] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowercase ( A_ , A_ )-> Union[str, Any]: '''simple docstring''' a : int = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) a : List[str] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_SCREAMING_SNAKE_CASE ) for lyr_num, lyr in enumerate(model.encoders ): a : Dict = weights[F'''layers_{lyr_num}'''] a : Tuple = ly_weight["attention"] a : int = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) a : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) a : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) a : Dict = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) a : Dict = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) a : int = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) a : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) a : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) a : int = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) a : Any = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowercase ( A_ , A_ )-> Dict: '''simple docstring''' a : int = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) a : Tuple = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) a : Optional[int] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_SCREAMING_SNAKE_CASE ) a : Tuple = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): a : List[str] = weights[F'''layers_{lyr_num}'''] a : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) a : Any = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) a : List[Any] = ly_weight["self_attention"] a : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) a : str = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) a : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) a : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) a : Union[str, Any] = ly_weight["MultiHeadDotProductAttention_0"] a : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) a : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) a : Any = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) a : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) a : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) a : Any = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) a : Optional[Any] = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) a : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) a : List[str] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) a : List[str] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) a : Any = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) a : str = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def lowercase ( A_ )-> List[str]: '''simple docstring''' a : Tuple = checkpoints.load_tax_checkpoint(args.checkpoint_path ) a : List[Any] = jnp.tree_util.tree_map(onp.array , _SCREAMING_SNAKE_CASE ) a : List[str] = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] a : str = os.path.join(args.checkpoint_path , ".." , "config.gin" ) a : List[Any] = inference.parse_training_gin_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a : Any = inference.InferenceModel(args.checkpoint_path , _SCREAMING_SNAKE_CASE ) a : int = DDPMScheduler(beta_schedule="squaredcos_cap_v2" , variance_type="fixed_large" ) a : Optional[Any] = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) a : Optional[Any] = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["targets_context"] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) a : List[Any] = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["targets_context"] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) a : List[str] = load_notes_encoder(ta_checkpoint["target"]["token_encoder"] , _SCREAMING_SNAKE_CASE ) a : List[str] = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"] , _SCREAMING_SNAKE_CASE ) a : Optional[int] = load_decoder(ta_checkpoint["target"]["decoder"] , _SCREAMING_SNAKE_CASE ) a : Optional[int] = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) a : Optional[Any] = SpectrogramDiffusionPipeline( notes_encoder=_SCREAMING_SNAKE_CASE , continuous_encoder=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE , melgan=_SCREAMING_SNAKE_CASE , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help="""Path to the original jax model checkpoint.""", ) __lowercase = parser.parse_args() main(args)
40
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _lowerCamelCase : Tuple = logging.get_logger(__name__) @add_end_docstrings(_a ) class __snake_case (_a ): def __init__( self : str , *_UpperCAmelCase : Union[str, Any] , **_UpperCAmelCase : Dict ) -> Any: '''simple docstring''' super().__init__(*_UpperCAmelCase , **_UpperCAmelCase ) self.check_model_type(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : Any=None , _UpperCAmelCase : Any=None , **_UpperCAmelCase : Dict ) -> int: '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Union[str, Any] = {}, {} if padding is not None: _lowerCAmelCase : Tuple = padding if truncation is not None: _lowerCAmelCase : List[str] = truncation if top_k is not None: _lowerCAmelCase : Union[str, Any] = top_k return preprocess_params, {}, postprocess_params def __call__( self : List[Any] , _UpperCAmelCase : Union["Image.Image", str] , _UpperCAmelCase : str = None , **_UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' if isinstance(_UpperCAmelCase , (Image.Image, str) ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ): _lowerCAmelCase : List[Any] = {"""image""": image, """question""": question} else: _lowerCAmelCase : Any = image _lowerCAmelCase : Tuple = super().__call__(_UpperCAmelCase , **_UpperCAmelCase ) return results def SCREAMING_SNAKE_CASE ( self : int , _UpperCAmelCase : str , _UpperCAmelCase : Dict=False , _UpperCAmelCase : Optional[int]=False ) -> List[Any]: '''simple docstring''' _lowerCAmelCase : str = load_image(inputs["""image"""] ) _lowerCAmelCase : Union[str, Any] = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=_UpperCAmelCase , truncation=_UpperCAmelCase ) _lowerCAmelCase : str = self.image_processor(images=_UpperCAmelCase , return_tensors=self.framework ) model_inputs.update(_UpperCAmelCase ) return model_inputs def SCREAMING_SNAKE_CASE ( self : Optional[int] , _UpperCAmelCase : Tuple ) -> str: '''simple docstring''' _lowerCAmelCase : str = self.model(**_UpperCAmelCase ) return model_outputs def SCREAMING_SNAKE_CASE ( self : Tuple , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any=5 ) -> List[Any]: '''simple docstring''' if top_k > self.model.config.num_labels: _lowerCAmelCase : List[Any] = self.model.config.num_labels if self.framework == "pt": _lowerCAmelCase : str = model_outputs.logits.sigmoid()[0] _lowerCAmelCase , _lowerCAmelCase : str = probs.topk(_UpperCAmelCase ) else: raise ValueError(f"Unsupported framework: {self.framework}" ) _lowerCAmelCase : Optional[int] = scores.tolist() _lowerCAmelCase : List[str] = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(_UpperCAmelCase , _UpperCAmelCase )]
159
import os from datetime import datetime as dt from github import Github _lowerCamelCase : List[Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def _UpperCAmelCase (): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = Github(os.environ["""GITHUB_TOKEN"""] ) _lowerCAmelCase : Any = g.get_repo("""huggingface/diffusers""" ) _lowerCAmelCase : Tuple = repo.get_issues(state="""open""" ) for issue in open_issues: _lowerCAmelCase : Tuple = sorted(issue.get_comments() , key=lambda UpperCamelCase_ : i.created_at , reverse=UpperCamelCase_ ) _lowerCAmelCase : List[Any] = comments[0] if len(UpperCamelCase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="""closed""" ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="""open""" ) issue.remove_from_labels("""stale""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) issue.add_to_labels("""stale""" ) if __name__ == "__main__": main()
159
1
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging _lowerCamelCase : str = logging.get_logger(__name__) logging.set_verbosity_info() def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> List[str]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: A__ = XLMProphetNetForConditionalGenerationOld.from_pretrained(lowercase_ ) A__ , A__ = XLMProphetNetForConditionalGeneration.from_pretrained( lowercase_ , output_loading_info=lowercase_ ) else: A__ = ProphetNetForConditionalGenerationOld.from_pretrained(lowercase_ ) A__ , A__ = ProphetNetForConditionalGeneration.from_pretrained( lowercase_ , output_loading_info=lowercase_ ) A__ = ['''key_proj''', '''value_proj''', '''query_proj'''] A__ = { '''self_attn''': '''ngram_self_attn''', '''cross_attn''': '''encoder_attn''', '''cross_attn_layer_norm''': '''encoder_attn_layer_norm''', '''feed_forward_layer_norm''': '''final_layer_norm''', '''feed_forward''': '''''', '''intermediate''': '''fc1''', '''output''': '''fc2''', '''key_proj''': '''k_proj''', '''query_proj''': '''q_proj''', '''value_proj''': '''v_proj''', '''word_embeddings''': '''embed_tokens''', '''embeddings_layer_norm''': '''emb_layer_norm''', '''relative_pos_embeddings''': '''relative_linear''', '''ngram_embeddings''': '''ngram_input_embed''', '''position_embeddings''': '''embed_positions''', } for key in loading_info["missing_keys"]: A__ = key.split('''.''' ) if attributes[0] == "lm_head": A__ = prophet A__ = prophet_old else: A__ = prophet.prophetnet A__ = prophet_old.model A__ = False for attribute in attributes: if attribute in mapping: A__ = mapping[attribute] if not hasattr(lowercase_ , lowercase_ ) and len(lowercase_ ) > 0: A__ = attribute elif hasattr(lowercase_ , lowercase_ ): A__ = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" A__ = old_model.weight logger.info(f"""{attribute} is initialized.""" ) A__ = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" A__ = old_model.bias logger.info(f"""{attribute} is initialized""" ) A__ = True break elif attribute in special_keys and hasattr(lowercase_ , '''in_proj_weight''' ): A__ = old_model.in_proj_weight.shape[0] // 3 A__ = getattr(lowercase_ , lowercase_ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": A__ = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) A__ = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": A__ = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) A__ = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": A__ = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) A__ = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) A__ = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." A__ = nn.Parameter(old_model.embed_positions.weight[:512, :] ) A__ = True break if attribute.isdigit(): A__ = model[int(lowercase_ )] A__ = old_model[int(lowercase_ )] else: A__ = getattr(lowercase_ , lowercase_ ) if old_attribute == "": A__ = old_model else: if not hasattr(lowercase_ , lowercase_ ): raise ValueError(f"""{old_model} does not have {old_attribute}""" ) A__ = getattr(lowercase_ , lowercase_ ) if not is_key_init: raise ValueError(f"""{key} was not correctly initialized!""" ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) prophet.save_pretrained(lowercase_ ) if __name__ == "__main__": _lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _lowerCamelCase : str = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
14
import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class UpperCamelCase_ : '''simple docstring''' def __init__( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int = 13 , UpperCAmelCase__ : int = 64 , UpperCAmelCase__ : int = 2 , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : int = 128 , UpperCAmelCase__ : Optional[Any]=[16, 32, 64, 128] , UpperCAmelCase__ : int = 7 , UpperCAmelCase__ : int = 4 , UpperCAmelCase__ : int = 37 , UpperCAmelCase__ : str = "gelu" , UpperCAmelCase__ : float = 0.1 , UpperCAmelCase__ : float = 0.1 , UpperCAmelCase__ : int = 10 , UpperCAmelCase__ : float = 0.02 , UpperCAmelCase__ : int = 2 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : int = 128 , UpperCAmelCase__ : List[int] = [2, 2, 2, 2] , UpperCAmelCase__ : int = 2 , UpperCAmelCase__ : int = 2 , ) ->List[Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = image_size A__ = patch_size A__ = num_channels A__ = is_training A__ = use_labels A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = type_sequence_label_size A__ = initializer_range A__ = encoder_stride A__ = num_attention_outputs A__ = embed_dim A__ = embed_dim + 1 A__ = resolution A__ = depths A__ = hidden_sizes A__ = dim A__ = mlp_expansion_ratio def SCREAMING_SNAKE_CASE ( self : List[Any]) ->str: '''simple docstring''' A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size) A__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : int) ->str: '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict) ->Dict: '''simple docstring''' A__ = TFEfficientFormerModel(config=UpperCAmelCase__) A__ = model(UpperCAmelCase__ , training=UpperCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : str) ->Union[str, Any]: '''simple docstring''' A__ = self.type_sequence_label_size A__ = TFEfficientFormerForImageClassification(UpperCAmelCase__) A__ = model(UpperCAmelCase__ , labels=UpperCAmelCase__ , training=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) # test greyscale images A__ = 1 A__ = TFEfficientFormerForImageClassification(UpperCAmelCase__) A__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) A__ = model(UpperCAmelCase__ , labels=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def SCREAMING_SNAKE_CASE ( self : int) ->List[str]: '''simple docstring''' A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) UpperCAmelCase__ = ( { '''feature-extraction''': TFEfficientFormerModel, '''image-classification''': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->List[str]: '''simple docstring''' A__ = TFEfficientFormerModelTester(self) A__ = ConfigTester( self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37) def SCREAMING_SNAKE_CASE ( self : int) ->Any: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''EfficientFormer does not use inputs_embeds''') def SCREAMING_SNAKE_CASE ( self : List[str]) ->Dict: '''simple docstring''' pass @unittest.skip(reason='''EfficientFormer does not support input and output embeddings''') def SCREAMING_SNAKE_CASE ( self : List[Any]) ->Optional[Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self : Any) ->Optional[Any]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(UpperCAmelCase__) A__ = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : str) ->Any: '''simple docstring''' def check_hidden_states_output(UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict): A__ = model_class(UpperCAmelCase__) A__ = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) , training=UpperCAmelCase__) A__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A__ = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1) self.assertEqual(len(UpperCAmelCase__) , UpperCAmelCase__) if hasattr(self.model_tester , '''encoder_seq_length'''): A__ = self.model_tester.encoder_seq_length if hasattr(self.model_tester , '''chunk_length''') and self.model_tester.chunk_length > 1: A__ = seq_length * self.model_tester.chunk_length else: A__ = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:]) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: A__ = outputs.decoder_hidden_states self.asseretIsInstance(UpperCAmelCase__ , (list, tuple)) self.assertEqual(len(UpperCAmelCase__) , UpperCAmelCase__) A__ = getattr(self.model_tester , '''seq_length''' , UpperCAmelCase__) A__ = getattr(self.model_tester , '''decoder_seq_length''' , UpperCAmelCase__) self.assertListEqual( list(hidden_states[-1].shape[-2:]) , [decoder_seq_length, self.model_tester.hidden_size] , ) A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict=False) ->int: '''simple docstring''' A__ = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Union[str, Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__) @unittest.skip(reason='''EfficientFormer does not implement masked image modeling yet''') def SCREAMING_SNAKE_CASE ( self : str) ->str: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Any) ->Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__) @slow def SCREAMING_SNAKE_CASE ( self : Tuple) ->Optional[int]: '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = TFEfficientFormerModel.from_pretrained(UpperCAmelCase__) self.assertIsNotNone(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Any) ->str: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = True A__ = getattr(self.model_tester , '''seq_length''' , UpperCAmelCase__) A__ = getattr(self.model_tester , '''encoder_seq_length''' , UpperCAmelCase__) A__ = getattr(self.model_tester , '''key_length''' , UpperCAmelCase__) A__ = getattr(self.model_tester , '''chunk_length''' , UpperCAmelCase__) if chunk_length is not None and hasattr(self.model_tester , '''num_hashes'''): A__ = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: A__ = True A__ = False A__ = True A__ = model_class(UpperCAmelCase__) A__ = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) , training=UpperCAmelCase__) A__ = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCAmelCase__) , self.model_tester.num_attention_outputs) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ = True A__ = model_class(UpperCAmelCase__) A__ = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) , training=UpperCAmelCase__) A__ = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCAmelCase__) , self.model_tester.num_attention_outputs) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:]) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def SCREAMING_SNAKE_CASE ( self : List[str]) ->Optional[Any]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model A__ = model_class(UpperCAmelCase__) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes A__ = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=UpperCAmelCase__) for key, val in model.input_signature.items() if key in model.dummy_inputs } A__ = model(UpperCAmelCase__) self.assertTrue(outputs_dict is not None) def SCREAMING_SNAKE_CASE ( ) -> Any: """simple docstring""" A__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE ( self : List[str]) ->List[str]: '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained('''snap-research/efficientformer-l1-300''') if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : List[str]) ->Any: '''simple docstring''' A__ = TFEfficientFormerForImageClassification.from_pretrained('''snap-research/efficientformer-l1-300''') A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=UpperCAmelCase__ , return_tensors='''tf''') # forward pass A__ = model(**UpperCAmelCase__ , training=UpperCAmelCase__) # verify the logits A__ = tf.TensorShape((1, 1_000)) self.assertEqual(outputs.logits.shape , UpperCAmelCase__) A__ = tf.constant([-0.0555, 0.4825, -0.0852]) self.assertTrue(np.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4)) @slow def SCREAMING_SNAKE_CASE ( self : Dict) ->int: '''simple docstring''' A__ = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( '''snap-research/efficientformer-l1-300''') A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=UpperCAmelCase__ , return_tensors='''tf''') # forward pass A__ = model(**UpperCAmelCase__ , training=UpperCAmelCase__) # verify the logits A__ = tf.TensorShape((1, 1_000)) self.assertEqual(outputs.logits.shape , UpperCAmelCase__) A__ = tf.constant([-0.1312, 0.4353, -1.0499]) self.assertTrue(np.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4))
14
1
"""simple docstring""" import math import random def lowercase_ ( _lowerCamelCase: float , _lowerCamelCase: bool = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __A = 0.02 def lowercase_ ( _lowerCamelCase: int , _lowerCamelCase: int ) -> float: '''simple docstring''' __lowerCamelCase : Union[str, Any] = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(snake_case__ ): # Forward propagation __lowerCamelCase : Dict = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? __lowerCamelCase : Any = (expected / 100) - layer_a # Error delta __lowerCamelCase : List[str] = layer_1_error * sigmoid_function(snake_case__ , snake_case__ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() __A = int(input('''Expected value: ''')) __A = int(input('''Number of propagations: ''')) print(forward_propagation(expected, number_propagations))
364
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def lowercase_ ( _lowerCamelCase: str , _lowerCamelCase: str ) -> str | Literal[False]: '''simple docstring''' __lowerCamelCase : Optional[int] = list(_lowerCamelCase ) __lowerCamelCase : Union[str, Any] = list(_lowerCamelCase ) __lowerCamelCase : Tuple = 0 for i in range(len(_lowerCamelCase ) ): if lista[i] != lista[i]: count += 1 __lowerCamelCase : Optional[int] = "_" if count > 1: return False else: return "".join(_lowerCamelCase ) def lowercase_ ( _lowerCamelCase: list[str] ) -> list[str]: '''simple docstring''' __lowerCamelCase : List[Any] = [] while True: __lowerCamelCase : Dict = ["$"] * len(_lowerCamelCase ) __lowerCamelCase : Any = [] for i in range(len(_lowerCamelCase ) ): for j in range(i + 1 , len(_lowerCamelCase ) ): __lowerCamelCase : str = compare_string(binary[i] , binary[j] ) if k is False: __lowerCamelCase : str = "*" __lowerCamelCase : Union[str, Any] = "*" temp.append("X" ) for i in range(len(_lowerCamelCase ) ): if checka[i] == "$": pi.append(binary[i] ) if len(_lowerCamelCase ) == 0: return pi __lowerCamelCase : Tuple = list(set(_lowerCamelCase ) ) def lowercase_ ( _lowerCamelCase: int , _lowerCamelCase: Sequence[float] ) -> list[str]: '''simple docstring''' __lowerCamelCase : Union[str, Any] = [] for minterm in minterms: __lowerCamelCase : Union[str, Any] = "" for _ in range(_lowerCamelCase ): __lowerCamelCase : Tuple = str(minterm % 2 ) + string minterm //= 2 temp.append(_lowerCamelCase ) return temp def lowercase_ ( _lowerCamelCase: str , _lowerCamelCase: str , _lowerCamelCase: int ) -> bool: '''simple docstring''' __lowerCamelCase : Tuple = list(_lowerCamelCase ) __lowerCamelCase : Optional[int] = list(_lowerCamelCase ) __lowerCamelCase : str = 0 for i in range(len(_lowerCamelCase ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def lowercase_ ( _lowerCamelCase: list[list[int]] , _lowerCamelCase: list[str] ) -> list[str]: '''simple docstring''' __lowerCamelCase : Optional[int] = [] __lowerCamelCase : str = [0] * len(_lowerCamelCase ) for i in range(len(chart[0] ) ): __lowerCamelCase : List[str] = 0 __lowerCamelCase : Optional[Any] = -1 for j in range(len(_lowerCamelCase ) ): if chart[j][i] == 1: count += 1 __lowerCamelCase : List[Any] = j if count == 1: __lowerCamelCase : Optional[Any] = 1 for i in range(len(_lowerCamelCase ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(_lowerCamelCase ) ): __lowerCamelCase : List[Any] = 0 temp.append(prime_implicants[i] ) while True: __lowerCamelCase : str = 0 __lowerCamelCase : Dict = -1 __lowerCamelCase : Tuple = 0 for i in range(len(_lowerCamelCase ) ): __lowerCamelCase : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: __lowerCamelCase : Optional[int] = count_n __lowerCamelCase : List[Any] = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(_lowerCamelCase ) ): __lowerCamelCase : Any = 0 def lowercase_ ( _lowerCamelCase: list[str] , _lowerCamelCase: list[str] ) -> list[list[int]]: '''simple docstring''' __lowerCamelCase : Dict = [[0 for x in range(len(_lowerCamelCase ) )] for x in range(len(_lowerCamelCase ) )] for i in range(len(_lowerCamelCase ) ): __lowerCamelCase : List[str] = prime_implicants[i].count("_" ) for j in range(len(_lowerCamelCase ) ): if is_for_table(prime_implicants[i] , binary[j] , _lowerCamelCase ): __lowerCamelCase : Dict = 1 return chart def lowercase_ ( ) -> None: '''simple docstring''' __lowerCamelCase : Any = int(input("Enter the no. of variables\n" ) ) __lowerCamelCase : List[str] = [ float(_lowerCamelCase ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] __lowerCamelCase : List[str] = decimal_to_binary(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase : str = check(_lowerCamelCase ) print("Prime Implicants are:" ) print(_lowerCamelCase ) __lowerCamelCase : Union[str, Any] = prime_implicant_chart(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase : Any = selection(_lowerCamelCase , _lowerCamelCase ) print("Essential Prime Implicants are:" ) print(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
0
import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO, ) a =logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[Any]: __lowerCamelCase : List[Any] = git.Repo(search_parent_directories=lowerCamelCase__ ) __lowerCamelCase : str = { 'repo_id': str(lowerCamelCase__ ), 'repo_sha': str(repo.head.object.hexsha ), 'repo_branch': str(repo.active_branch ), } with open(os.path.join(lowerCamelCase__ , 'git_log.json' ) , 'w' ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ , indent=4 ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Any: if params.n_gpu <= 0: __lowerCamelCase : Any = 0 __lowerCamelCase : Union[str, Any] = -1 __lowerCamelCase : str = True __lowerCamelCase : List[Any] = False return assert torch.cuda.is_available() logger.info('Initializing GPUs' ) if params.n_gpu > 1: assert params.local_rank != -1 __lowerCamelCase : Tuple = int(os.environ['WORLD_SIZE'] ) __lowerCamelCase : int = int(os.environ['N_GPU_NODE'] ) __lowerCamelCase : Dict = int(os.environ['RANK'] ) # number of nodes / node ID __lowerCamelCase : Dict = params.world_size // params.n_gpu_per_node __lowerCamelCase : int = params.global_rank // params.n_gpu_per_node __lowerCamelCase : Union[str, Any] = True assert params.n_nodes == int(os.environ['N_NODES'] ) assert params.node_id == int(os.environ['NODE_RANK'] ) # local job (single GPU) else: assert params.local_rank == -1 __lowerCamelCase : int = 1 __lowerCamelCase : Dict = 0 __lowerCamelCase : Tuple = 0 __lowerCamelCase : Dict = 0 __lowerCamelCase : Union[str, Any] = 1 __lowerCamelCase : Optional[Any] = 1 __lowerCamelCase : Dict = False # sanity checks assert params.n_nodes >= 1 assert 0 <= params.node_id < params.n_nodes assert 0 <= params.local_rank <= params.global_rank < params.world_size assert params.world_size == params.n_nodes * params.n_gpu_per_node # define whether this is the master process / if we are in multi-node distributed mode __lowerCamelCase : List[str] = params.node_id == 0 and params.local_rank == 0 __lowerCamelCase : Dict = params.n_nodes > 1 # summary __lowerCamelCase : List[str] = F"--- Global rank: {params.global_rank} - " logger.info(PREFIX + 'Number of nodes: %i' % params.n_nodes ) logger.info(PREFIX + 'Node ID : %i' % params.node_id ) logger.info(PREFIX + 'Local rank : %i' % params.local_rank ) logger.info(PREFIX + 'World size : %i' % params.world_size ) logger.info(PREFIX + 'GPUs per node : %i' % params.n_gpu_per_node ) logger.info(PREFIX + 'Master : %s' % str(params.is_master ) ) logger.info(PREFIX + 'Multi-node : %s' % str(params.multi_node ) ) logger.info(PREFIX + 'Multi-GPU : %s' % str(params.multi_gpu ) ) logger.info(PREFIX + 'Hostname : %s' % socket.gethostname() ) # set GPU device torch.cuda.set_device(params.local_rank ) # initialize multi-GPU if params.multi_gpu: logger.info('Initializing PyTorch distributed' ) torch.distributed.init_process_group( init_method='env://' , backend='nccl' , ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Union[str, Any]: np.random.seed(args.seed ) torch.manual_seed(args.seed ) if args.n_gpu > 0: torch.cuda.manual_seed_all(args.seed )
73
from bisect import bisect from itertools import accumulate def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Union[str, Any]: __lowerCamelCase : Optional[Any] = sorted(zip(lowerCamelCase__ , lowerCamelCase__ ) , key=lambda lowerCamelCase__ : x[0] / x[1] , reverse=lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Any = [i[0] for i in r], [i[1] for i in r] __lowerCamelCase : List[str] = list(accumulate(lowerCamelCase__ ) ) __lowerCamelCase : Union[str, Any] = bisect(lowerCamelCase__ , lowerCamelCase__ ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
73
1
'''simple docstring''' from torch import nn def _lowerCAmelCase ( __snake_case : str ) -> List[Any]: if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'Unsupported activation function: {act_fn}' )
190
'''simple docstring''' from __future__ import annotations from math import gcd def _lowerCAmelCase ( __snake_case : int , __snake_case : int = 2 , __snake_case : int = 1 , __snake_case : int = 3 , ) -> int | None: # A value less than 2 can cause an infinite loop in the algorithm. if num < 2: raise ValueError('The input value cannot be less than 2' ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(__snake_case : int , __snake_case : int , __snake_case : int ) -> int: return (pow(__snake_case , 2 ) + step) % modulus for _ in range(__snake_case ): # These track the position within the cycle detection logic. __A : int = seed __A : Union[str, Any] = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. __A : List[Any] = rand_fn(__snake_case , __snake_case , __snake_case ) __A : Optional[Any] = rand_fn(__snake_case , __snake_case , __snake_case ) __A : Any = rand_fn(__snake_case , __snake_case , __snake_case ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. __A : Optional[int] = gcd(hare - tortoise , __snake_case ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. __A : Union[str, Any] = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse lowercase__ : str = argparse.ArgumentParser() parser.add_argument( '''num''', type=int, help='''The value to find a divisor of''', ) parser.add_argument( '''--attempts''', type=int, default=3, help='''The number of attempts before giving up''', ) lowercase__ : Optional[int] = parser.parse_args() lowercase__ : int = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(f"""{args.num} is probably prime""") else: lowercase__ : List[str] = args.num // divisor print(f"""{args.num} = {divisor} * {quotient}""")
190
1
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def _a ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" if isinstance(SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(SCREAMING_SNAKE_CASE , PIL.Image.Image ): UpperCamelCase__ : str = [image] if isinstance(image[0] , PIL.Image.Image ): UpperCamelCase__ : str = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] UpperCamelCase__ : Optional[Any] = np.concatenate(SCREAMING_SNAKE_CASE , axis=0 ) UpperCamelCase__ : Any = np.array(SCREAMING_SNAKE_CASE ).astype(np.floataa ) / 255.0 UpperCamelCase__ : Tuple = image.transpose(0 , 3 , 1 , 2 ) UpperCamelCase__ : Optional[int] = 2.0 * image - 1.0 UpperCamelCase__ : Tuple = torch.from_numpy(SCREAMING_SNAKE_CASE ) elif isinstance(image[0] , torch.Tensor ): UpperCamelCase__ : Dict = torch.cat(SCREAMING_SNAKE_CASE , dim=0 ) return image def _a ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : List[Any]=0.9995 ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ): UpperCamelCase__ : str = True UpperCamelCase__ : Optional[Any] = va.device UpperCamelCase__ : Dict = va.cpu().numpy() UpperCamelCase__ : Dict = va.cpu().numpy() UpperCamelCase__ : Optional[int] = np.sum(va * va / (np.linalg.norm(SCREAMING_SNAKE_CASE ) * np.linalg.norm(SCREAMING_SNAKE_CASE )) ) if np.abs(SCREAMING_SNAKE_CASE ) > DOT_THRESHOLD: UpperCamelCase__ : Union[str, Any] = (1 - t) * va + t * va else: UpperCamelCase__ : List[str] = np.arccos(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Any = np.sin(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Union[str, Any] = theta_a * t UpperCamelCase__ : str = np.sin(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[Any] = np.sin(theta_a - theta_t ) / sin_theta_a UpperCamelCase__ : Optional[int] = sin_theta_t / sin_theta_a UpperCamelCase__ : Optional[Any] = sa * va + sa * va if inputs_are_torch: UpperCamelCase__ : Any = torch.from_numpy(SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) return va def _a ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : str ): """simple docstring""" UpperCamelCase__ : List[Any] = F.normalize(SCREAMING_SNAKE_CASE , dim=-1 ) UpperCamelCase__ : Any = F.normalize(SCREAMING_SNAKE_CASE , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def _a ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" for param in model.parameters(): UpperCamelCase__ : Optional[int] = value class __magic_name__ ( __lowerCAmelCase): def __init__( self : List[str] , lowerCamelCase__ : AutoencoderKL , lowerCamelCase__ : CLIPTextModel , lowerCamelCase__ : CLIPModel , lowerCamelCase__ : CLIPTokenizer , lowerCamelCase__ : UNetaDConditionModel , lowerCamelCase__ : Union[PNDMScheduler, LMSDiscreteScheduler, DDIMScheduler, DPMSolverMultistepScheduler] , lowerCamelCase__ : CLIPFeatureExtractor , lowerCamelCase__ : Optional[int]=None , lowerCamelCase__ : Any=None , lowerCamelCase__ : Any=None , ) -> Union[str, Any]: '''simple docstring''' super().__init__() self.register_modules( vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , clip_model=lowerCamelCase__ , tokenizer=lowerCamelCase__ , unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , coca_model=lowerCamelCase__ , coca_tokenizer=lowerCamelCase__ , coca_transform=lowerCamelCase__ , ) UpperCamelCase__ : Optional[int] = ( feature_extractor.size if isinstance(feature_extractor.size , lowerCamelCase__ ) else feature_extractor.size['''shortest_edge'''] ) UpperCamelCase__ : str = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , lowerCamelCase__ ) set_requires_grad(self.clip_model , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Optional[Union[str, int]] = "auto" ) -> Tuple: '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCamelCase__ : Dict = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' self.enable_attention_slicing(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Any ) -> Optional[int]: '''simple docstring''' set_requires_grad(self.vae , lowerCamelCase__ ) def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' set_requires_grad(self.vae , lowerCamelCase__ ) def UpperCAmelCase__ ( self : str ) -> str: '''simple docstring''' set_requires_grad(self.unet , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' set_requires_grad(self.unet , lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : str ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ : Optional[Any] = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) UpperCamelCase__ : List[Any] = max(num_inference_steps - init_timestep , 0 ) UpperCamelCase__ : Optional[Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : str , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Union[str, Any]=None ) -> Any: '''simple docstring''' if not isinstance(lowerCamelCase__ , torch.Tensor ): raise ValueError(F"`image` has to be of type `torch.Tensor` but is {type(lowerCamelCase__ )}" ) UpperCamelCase__ : List[Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase__ : Tuple = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(lowerCamelCase__ ) ] UpperCamelCase__ : Dict = torch.cat(lowerCamelCase__ , dim=0 ) else: UpperCamelCase__ : int = self.vae.encode(lowerCamelCase__ ).latent_dist.sample(lowerCamelCase__ ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCamelCase__ : Union[str, Any] = 0.1_8215 * init_latents UpperCamelCase__ : Union[str, Any] = init_latents.repeat_interleave(lowerCamelCase__ , dim=0 ) UpperCamelCase__ : Dict = randn_tensor(init_latents.shape , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents UpperCamelCase__ : str = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Tuple = init_latents return latents def UpperCAmelCase__ ( self : List[str] , lowerCamelCase__ : Optional[Any] ) -> List[str]: '''simple docstring''' UpperCamelCase__ : Optional[int] = self.coca_transform(lowerCamelCase__ ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): UpperCamelCase__ : Optional[int] = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) UpperCamelCase__ : int = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split('''<end_of_text>''' )[0].replace('''<start_of_text>''' , '''''' ).rstrip(''' .,''' ) def UpperCAmelCase__ ( self : int , lowerCamelCase__ : str , lowerCamelCase__ : str ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = self.feature_extractor.preprocess(lowerCamelCase__ ) UpperCamelCase__ : Dict = torch.from_numpy(clip_image_input['''pixel_values'''][0] ).unsqueeze(0 ).to(self.device ).half() UpperCamelCase__ : Union[str, Any] = self.clip_model.get_image_features(lowerCamelCase__ ) UpperCamelCase__ : Dict = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = image_embeddings_clip.repeat_interleave(lowerCamelCase__ , dim=0 ) return image_embeddings_clip @torch.enable_grad() def UpperCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : str , lowerCamelCase__ : Optional[Any] , ) -> List[str]: '''simple docstring''' UpperCamelCase__ : str = latents.detach().requires_grad_() UpperCamelCase__ : Tuple = self.scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) # predict the noise residual UpperCamelCase__ : Tuple = self.unet(lowerCamelCase__ , lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): UpperCamelCase__ : List[Any] = self.scheduler.alphas_cumprod[timestep] UpperCamelCase__ : List[Any] = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCamelCase__ : Optional[Any] = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 UpperCamelCase__ : Optional[Any] = torch.sqrt(lowerCamelCase__ ) UpperCamelCase__ : str = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , lowerCamelCase__ ): UpperCamelCase__ : List[Any] = self.scheduler.sigmas[index] UpperCamelCase__ : List[Any] = latents - sigma * noise_pred else: raise ValueError(F"scheduler type {type(self.scheduler )} not supported" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCamelCase__ : Tuple = 1 / 0.1_8215 * sample UpperCamelCase__ : List[str] = self.vae.decode(lowerCamelCase__ ).sample UpperCamelCase__ : Any = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase__ : str = transforms.Resize(self.feature_extractor_size )(lowerCamelCase__ ) UpperCamelCase__ : Tuple = self.normalize(lowerCamelCase__ ).to(latents.dtype ) UpperCamelCase__ : int = self.clip_model.get_image_features(lowerCamelCase__ ) UpperCamelCase__ : List[Any] = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=lowerCamelCase__ ) UpperCamelCase__ : Any = spherical_dist_loss(lowerCamelCase__ , lowerCamelCase__ ).mean() * clip_guidance_scale UpperCamelCase__ : Dict = -torch.autograd.grad(lowerCamelCase__ , lowerCamelCase__ )[0] if isinstance(self.scheduler , lowerCamelCase__ ): UpperCamelCase__ : Optional[Any] = latents.detach() + grads * (sigma**2) UpperCamelCase__ : Union[str, Any] = noise_pred_original else: UpperCamelCase__ : List[Any] = noise_pred_original - torch.sqrt(lowerCamelCase__ ) * grads return noise_pred, latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase__ : Optional[str] = None , lowerCamelCase__ : Optional[str] = None , lowerCamelCase__ : Optional[int] = 512 , lowerCamelCase__ : Optional[int] = 512 , lowerCamelCase__ : float = 0.6 , lowerCamelCase__ : Optional[int] = 50 , lowerCamelCase__ : Optional[float] = 7.5 , lowerCamelCase__ : Optional[int] = 1 , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : Optional[float] = 100 , lowerCamelCase__ : Optional[torch.Generator] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : float = 0.1 , lowerCamelCase__ : float = 0.1 , ) -> Optional[int]: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError(F"You have passed {batch_size} batch_size, but only {len(lowerCamelCase__ )} generators." ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if isinstance(lowerCamelCase__ , torch.Generator ) and batch_size > 1: UpperCamelCase__ : List[Any] = [generator] + [None] * (batch_size - 1) UpperCamelCase__ : str = [ ('''model''', self.coca_model is None), ('''tokenizer''', self.coca_tokenizer is None), ('''transform''', self.coca_transform is None), ] UpperCamelCase__ : Dict = [x[0] for x in coca_is_none if x[1]] UpperCamelCase__ : Optional[int] = ''', '''.join(lowerCamelCase__ ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(lowerCamelCase__ ): raise ValueError( F"Content prompt is None and CoCa [{coca_is_none_str}] is None." F"Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) UpperCamelCase__ : str = self.get_image_description(lowerCamelCase__ ) if style_prompt is None: if len(lowerCamelCase__ ): raise ValueError( F"Style prompt is None and CoCa [{coca_is_none_str}] is None." F" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) UpperCamelCase__ : Optional[Any] = self.get_image_description(lowerCamelCase__ ) # get prompt text embeddings for content and style UpperCamelCase__ : int = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) UpperCamelCase__ : str = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] UpperCamelCase__ : int = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) UpperCamelCase__ : List[str] = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] UpperCamelCase__ : Any = slerp(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # duplicate text embeddings for each generation per prompt UpperCamelCase__ : Any = text_embeddings.repeat_interleave(lowerCamelCase__ , dim=0 ) # set timesteps UpperCamelCase__ : List[str] = '''offset''' in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) UpperCamelCase__ : Dict = {} if accepts_offset: UpperCamelCase__ : Union[str, Any] = 1 self.scheduler.set_timesteps(lowerCamelCase__ , **lowerCamelCase__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) UpperCamelCase__ , UpperCamelCase__ : List[str] = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) UpperCamelCase__ : Tuple = timesteps[:1].repeat(lowerCamelCase__ ) # Preprocess image UpperCamelCase__ : Optional[int] = preprocess(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Dict = self.prepare_latents( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , text_embeddings.dtype , self.device , lowerCamelCase__ ) UpperCamelCase__ : Tuple = preprocess(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : List[Any] = self.prepare_latents( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , text_embeddings.dtype , self.device , lowerCamelCase__ ) UpperCamelCase__ : Dict = slerp(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if clip_guidance_scale > 0: UpperCamelCase__ : Tuple = self.get_clip_image_embeddings(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = self.get_clip_image_embeddings(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Tuple = slerp( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. UpperCamelCase__ : Union[str, Any] = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: UpperCamelCase__ : str = content_text_input.input_ids.shape[-1] UpperCamelCase__ : Tuple = self.tokenizer([''''''] , padding='''max_length''' , max_length=lowerCamelCase__ , return_tensors='''pt''' ) UpperCamelCase__ : Tuple = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt UpperCamelCase__ : Optional[int] = uncond_embeddings.repeat_interleave(lowerCamelCase__ , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCamelCase__ : Any = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. UpperCamelCase__ : Dict = (batch_size, self.unet.config.in_channels, height // 8, width // 8) UpperCamelCase__ : Tuple = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps UpperCamelCase__ : Tuple = torch.randn(lowerCamelCase__ , generator=lowerCamelCase__ , device='''cpu''' , dtype=lowerCamelCase__ ).to( self.device ) else: UpperCamelCase__ : List[str] = torch.randn(lowerCamelCase__ , generator=lowerCamelCase__ , device=self.device , dtype=lowerCamelCase__ ) else: if latents.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) UpperCamelCase__ : Tuple = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler UpperCamelCase__ : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCamelCase__ : Optional[int] = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCamelCase__ : List[Any] = {} if accepts_eta: UpperCamelCase__ : int = eta # check if the scheduler accepts generator UpperCamelCase__ : str = '''generator''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: UpperCamelCase__ : str = generator with self.progress_bar(total=lowerCamelCase__ ): for i, t in enumerate(lowerCamelCase__ ): # expand the latents if we are doing classifier free guidance UpperCamelCase__ : Optional[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCamelCase__ : List[str] = self.scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) # predict the noise residual UpperCamelCase__ : Optional[Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ ).sample # perform classifier free guidance if do_classifier_free_guidance: UpperCamelCase__ , UpperCamelCase__ : Tuple = noise_pred.chunk(2 ) UpperCamelCase__ : Optional[int] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: UpperCamelCase__ : List[Any] = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) UpperCamelCase__ , UpperCamelCase__ : int = self.cond_fn( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase__ : int = self.scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCamelCase__ : List[str] = 1 / 0.1_8215 * latents UpperCamelCase__ : int = self.vae.decode(lowerCamelCase__ ).sample UpperCamelCase__ : List[Any] = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase__ : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase__ : Any = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=lowerCamelCase__ , nsfw_content_detected=lowerCamelCase__ )
146
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class __magic_name__ ( __lowerCAmelCase , unittest.TestCase): A: str = XLMTokenizer A: Optional[Any] = False def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase__ : Any = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] UpperCamelCase__ : Optional[int] = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) UpperCamelCase__ : Optional[Any] = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] UpperCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) UpperCamelCase__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(lowerCamelCase__ ) ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Dict ) -> Tuple: '''simple docstring''' UpperCamelCase__ : int = '''lower newer''' UpperCamelCase__ : List[str] = '''lower newer''' return input_text, output_text def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) UpperCamelCase__ : Tuple = '''lower''' UpperCamelCase__ : Dict = ['''low''', '''er</w>'''] UpperCamelCase__ : Optional[int] = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) UpperCamelCase__ : Dict = tokens + ['''<unk>'''] UpperCamelCase__ : List[Any] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) @slow def UpperCAmelCase__ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Any = XLMTokenizer.from_pretrained('''xlm-mlm-en-2048''' ) UpperCamelCase__ : List[str] = tokenizer.encode('''sequence builders''' , add_special_tokens=lowerCamelCase__ ) UpperCamelCase__ : Any = tokenizer.encode('''multi-sequence build''' , add_special_tokens=lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) UpperCamelCase__ : Any = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ , lowerCamelCase__ ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
146
1
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCAmelCase_ ( _a, unittest.TestCase): '''simple docstring''' __UpperCamelCase : Optional[int] = DDIMPipeline __UpperCamelCase : Any = UNCONDITIONAL_IMAGE_GENERATION_PARAMS __UpperCamelCase : int = PipelineTesterMixin.required_optional_params - { "num_images_per_prompt", "latents", "callback", "callback_steps", } __UpperCamelCase : str = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS __UpperCamelCase : List[str] = False def _lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) UpperCamelCase : Dict = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) UpperCamelCase : List[str] = DDIMScheduler() UpperCamelCase : str = {'''unet''': unet, '''scheduler''': scheduler} return components def _lowercase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ): """simple docstring""" if str(__SCREAMING_SNAKE_CASE ).startswith('''mps''' ): UpperCamelCase : List[str] = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: UpperCamelCase : List[str] = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) UpperCamelCase : Any = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def _lowercase ( self ): """simple docstring""" UpperCamelCase : Dict = '''cpu''' UpperCamelCase : Any = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) UpperCamelCase : Any = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[Any] = pipe(**__SCREAMING_SNAKE_CASE ).images UpperCamelCase : int = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) UpperCamelCase : Tuple = np.array( [1.000e00, 5.717e-01, 4.717e-01, 1.000e00, 0.000e00, 1.000e00, 3.000e-04, 0.000e00, 9.000e-04] ) UpperCamelCase : Any = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__SCREAMING_SNAKE_CASE , 1e-3 ) def _lowercase ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _lowercase ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=3e-3 ) def _lowercase ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3e-3 ) def _lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase): '''simple docstring''' def _lowercase ( self ): """simple docstring""" UpperCamelCase : Dict = '''google/ddpm-cifar10-32''' UpperCamelCase : str = UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) UpperCamelCase : List[str] = DDIMScheduler() UpperCamelCase : List[Any] = DDIMPipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) ddim.to(__SCREAMING_SNAKE_CASE ) ddim.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) UpperCamelCase : str = torch.manual_seed(0 ) UpperCamelCase : List[Any] = ddim(generator=__SCREAMING_SNAKE_CASE , eta=0.0 , output_type='''numpy''' ).images UpperCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCamelCase : str = np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ): """simple docstring""" UpperCamelCase : Union[str, Any] = '''google/ddpm-ema-bedroom-256''' UpperCamelCase : Optional[Any] = UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = DDIMScheduler.from_pretrained(__SCREAMING_SNAKE_CASE ) UpperCamelCase : int = DDIMPipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) ddpm.to(__SCREAMING_SNAKE_CASE ) ddpm.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = torch.manual_seed(0 ) UpperCamelCase : List[str] = ddpm(generator=__SCREAMING_SNAKE_CASE , output_type='''numpy''' ).images UpperCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) UpperCamelCase : Optional[int] = np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
361
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def a ( SCREAMING_SNAKE_CASE_ : dict ): """simple docstring""" return (data["data"], data["target"]) def a ( SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : np.ndarray ): """simple docstring""" UpperCamelCase : Optional[Any] = XGBRegressor(verbosity=0 , random_state=4_2 ) xgb.fit(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Predict target for test data UpperCamelCase : Any = xgb.predict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = predictions.reshape(len(SCREAMING_SNAKE_CASE_ ) , 1 ) return predictions def a ( ): """simple docstring""" UpperCamelCase : Tuple = fetch_california_housing() UpperCamelCase , UpperCamelCase : Tuple = data_handling(SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = train_test_split( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , test_size=0.25 , random_state=1 ) UpperCamelCase : Optional[Any] = xgboost(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Error printing print(F"""Mean Absolute Error : {mean_absolute_error(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}""" ) print(F"""Mean Square Error : {mean_squared_error(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}""" ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
315
0
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ) -> int: return abs(UpperCAmelCase_ ) if a == 0 else greatest_common_divisor(b % a , UpperCAmelCase_ ) def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. __lowerCamelCase , __lowerCamelCase : Optional[Any] = y, x % y return abs(UpperCAmelCase_ ) def UpperCAmelCase__ ( ) -> List[str]: try: __lowerCamelCase : List[Any] = input('Enter two integers separated by comma (,): ' ).split(',' ) __lowerCamelCase : Optional[int] = int(nums[0] ) __lowerCamelCase : List[Any] = int(nums[1] ) print( F'greatest_common_divisor({num_a}, {num_a}) = ' F'{greatest_common_divisor(UpperCAmelCase_ , UpperCAmelCase_ )}' ) print(F'By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCAmelCase_ , UpperCAmelCase_ )}' ) except (IndexError, UnboundLocalError, ValueError): print('Wrong input' ) if __name__ == "__main__": main()
185
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[Any] = { """configuration_megatron_bert""": ["""MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MegatronBertConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : int = [ """MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MegatronBertForCausalLM""", """MegatronBertForMaskedLM""", """MegatronBertForMultipleChoice""", """MegatronBertForNextSentencePrediction""", """MegatronBertForPreTraining""", """MegatronBertForQuestionAnswering""", """MegatronBertForSequenceClassification""", """MegatronBertForTokenClassification""", """MegatronBertModel""", """MegatronBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys A__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
185
1
'''simple docstring''' import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) __lowerCAmelCase = logging.getLogger() def __lowerCamelCase ( ) -> int: _a : Tuple = argparse.ArgumentParser() parser.add_argument('-f' ) _a : Optional[int] = parser.parse_args() return args.f def __lowerCamelCase ( lowerCAmelCase_ ) -> Union[str, Any]: _a : Union[str, Any] = {} _a : Dict = os.path.join(lowerCAmelCase_ , 'all_results.json' ) if os.path.exists(lowerCAmelCase_ ): with open(lowerCAmelCase_ , 'r' ) as f: _a : str = json.load(lowerCAmelCase_ ) else: raise ValueError(f"""can't find {path}""" ) return results def __lowerCamelCase ( ) -> List[str]: _a : List[Any] = torch.cuda.is_available() and torch_device == 'cuda' return is_using_cuda and is_apex_available() __lowerCAmelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __magic_name__ ( _UpperCamelCase ): @classmethod def __lowercase ( cls : List[str] ): # Write Accelerate config, will pick up on CPU, GPU, and multi-GPU _a : int = tempfile.mkdtemp() _a : List[Any] = os.path.join(cls.tmpdir ,'default_config.yml' ) write_basic_config(save_location=cls.configPath ) _a : Union[str, Any] = ['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def __lowercase ( cls : Tuple ): shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Optional[Any] ): _a : List[Any] = self.get_auto_remove_tmp_dir() _a : Any = F""" {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --seed=42 --checkpointing_steps epoch --with_tracking """.split() if is_cuda_and_apex_available(): testargs.append('--fp16' ) run_command(self._launch_args + testargs ) _a : List[str] = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_accuracy'] ,0.75 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'glue_no_trainer' ) ) ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Optional[Any] ): _a : Optional[int] = self.get_auto_remove_tmp_dir() _a : Any = F""" {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --block_size 128 --per_device_train_batch_size 5 --per_device_eval_batch_size 5 --num_train_epochs 2 --output_dir {tmp_dir} --checkpointing_steps epoch --with_tracking """.split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) _a : Optional[Any] = get_results(_UpperCAmelCase ) self.assertLess(result['perplexity'] ,100 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'clm_no_trainer' ) ) ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Dict ): _a : List[Any] = self.get_auto_remove_tmp_dir() _a : Dict = F""" {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --num_train_epochs=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _a : List[Any] = get_results(_UpperCAmelCase ) self.assertLess(result['perplexity'] ,42 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'mlm_no_trainer' ) ) ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Union[str, Any] ): # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu _a : Optional[int] = 7 if get_gpu_count() > 1 else 2 _a : Tuple = self.get_auto_remove_tmp_dir() _a : Any = F""" {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _a : Any = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_accuracy'] ,0.75 ) self.assertLess(result['train_loss'] ,0.5 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'ner_no_trainer' ) ) ) @unittest.skip(reason='Fix me @muellerzr' ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : int ): _a : List[Any] = self.get_auto_remove_tmp_dir() _a : int = F""" {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --seed=42 --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _a : List[str] = get_results(_UpperCAmelCase ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result['eval_f1'] ,28 ) self.assertGreaterEqual(result['eval_exact'] ,28 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'qa_no_trainer' ) ) ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Tuple ): _a : Optional[int] = self.get_auto_remove_tmp_dir() _a : List[str] = F""" {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/swag/sample.json --validation_file tests/fixtures/tests_samples/swag/sample.json --output_dir {tmp_dir} --max_train_steps=20 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --with_tracking """.split() run_command(self._launch_args + testargs ) _a : Tuple = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_accuracy'] ,0.8 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'swag_no_trainer' ) ) ) @slow @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : str ): _a : List[Any] = self.get_auto_remove_tmp_dir() _a : List[str] = F""" {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _a : Tuple = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_rouge1'] ,10 ) self.assertGreaterEqual(result['eval_rouge2'] ,2 ) self.assertGreaterEqual(result['eval_rougeL'] ,7 ) self.assertGreaterEqual(result['eval_rougeLsum'] ,7 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'summarization_no_trainer' ) ) ) @slow @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : List[str] ): _a : Any = self.get_auto_remove_tmp_dir() _a : Optional[Any] = F""" {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py --model_name_or_path sshleifer/student_marian_en_ro_6_1 --source_lang en --target_lang ro --train_file tests/fixtures/tests_samples/wmt16/sample.json --validation_file tests/fixtures/tests_samples/wmt16/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --num_beams=6 --learning_rate=3e-3 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --source_lang en_XX --target_lang ro_RO --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _a : List[str] = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_bleu'] ,30 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'epoch_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'translation_no_trainer' ) ) ) @slow def __lowercase ( self : Any ): _a : Optional[int] = logging.StreamHandler(sys.stdout ) logger.addHandler(_UpperCAmelCase ) _a : List[str] = self.get_auto_remove_tmp_dir() _a : List[str] = F""" {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py --dataset_name huggingface/semantic-segmentation-test-sample --output_dir {tmp_dir} --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch """.split() run_command(self._launch_args + testargs ) _a : int = get_results(_UpperCAmelCase ) self.assertGreaterEqual(result['eval_overall_accuracy'] ,0.10 ) @mock.patch.dict(os.environ ,{'WANDB_MODE': 'offline'} ) def __lowercase ( self : Any ): _a : Optional[Any] = self.get_auto_remove_tmp_dir() _a : List[str] = F""" {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py --model_name_or_path google/vit-base-patch16-224-in21k --dataset_name hf-internal-testing/cats_vs_dogs_sample --learning_rate 1e-4 --per_device_train_batch_size 2 --per_device_eval_batch_size 1 --max_train_steps 2 --train_val_split 0.1 --seed 42 --output_dir {tmp_dir} --with_tracking --checkpointing_steps 1 """.split() if is_cuda_and_apex_available(): testargs.append('--fp16' ) run_command(self._launch_args + testargs ) _a : Any = get_results(_UpperCAmelCase ) # The base model scores a 25% self.assertGreaterEqual(result['eval_accuracy'] ,0.6 ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'step_1' ) ) ) self.assertTrue(os.path.exists(os.path.join(_UpperCAmelCase ,'image_classification_no_trainer' ) ) )
354
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __lowerCAmelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __lowerCAmelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __lowerCAmelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1_000)) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> tuple[str, float]: _a : List[Any] = len([g for position, g in enumerate(lowerCAmelCase_ ) if g == main_target[position]] ) return (item, float(lowerCAmelCase_ )) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> tuple[str, str]: _a : Dict = random.randint(0 , len(lowerCAmelCase_ ) - 1 ) _a : Optional[int] = parent_a[:random_slice] + parent_a[random_slice:] _a : Optional[int] = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _a : Optional[Any] = list(lowerCAmelCase_ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _a : Optional[int] = random.choice(lowerCAmelCase_ ) return "".join(lowerCAmelCase_ ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) -> list[str]: _a : List[str] = [] # Generate more children proportionally to the fitness score. _a : Tuple = int(parent_a[1] * 100 ) + 1 _a : Tuple = 10 if child_n >= 10 else child_n for _ in range(lowerCAmelCase_ ): _a : Any = population_score[random.randint(0 , lowerCAmelCase_ )][0] _a , _a : Tuple = crossover(parent_a[0] , lowerCAmelCase_ ) # Append new string to the population list. pop.append(mutate(lowerCAmelCase_ , lowerCAmelCase_ ) ) pop.append(mutate(lowerCAmelCase_ , lowerCAmelCase_ ) ) return pop def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _a : Dict = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(lowerCAmelCase_ ) # Verify that the target contains no genes besides the ones inside genes variable. _a : Optional[int] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _a : List[Any] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(lowerCAmelCase_ ) # Generate random starting population. _a : Union[str, Any] = [] for _ in range(lowerCAmelCase_ ): population.append(''.join([random.choice(lowerCAmelCase_ ) for i in range(len(lowerCAmelCase_ ) )] ) ) # Just some logs to know what the algorithms is doing. _a , _a : Union[str, Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowerCAmelCase_ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _a : Optional[Any] = [evaluate(lowerCAmelCase_ , lowerCAmelCase_ ) for item in population] # Check if there is a matching evolution. _a : Tuple = sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x[1] , reverse=lowerCAmelCase_ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _a : Dict = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowerCAmelCase_ ) # Normalize population score to be between 0 and 1. _a : Tuple = [ (item, score / len(lowerCAmelCase_ )) for item, score in population_score ] # This is selection for i in range(lowerCAmelCase_ ): population.extend(select(population_score[int(lowerCAmelCase_ )] , lowerCAmelCase_ , lowerCAmelCase_ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowerCAmelCase_ ) > N_POPULATION: break if __name__ == "__main__": __lowerCAmelCase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __lowerCAmelCase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = basic(target_str, genes_list) print( f"""\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}""" )
107
0
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" lowerCamelCase__ = inspect.getfile(accelerate.test_utils ) lowerCamelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_cli.py"""] ) lowerCamelCase__ = ["""accelerate""", """launch"""] lowerCamelCase__ = Path.home() / """.cache/huggingface/accelerate""" lowerCamelCase__ = """default_config.yaml""" lowerCamelCase__ = config_folder / config_file lowerCamelCase__ = config_folder / """_default_config.yaml""" lowerCamelCase__ = Path("""tests/test_configs""" ) @classmethod def A ( cls : int ) -> Any: if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def A ( cls : Dict ) -> Optional[int]: if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def A ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase : Tuple = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def A ( self : Any ) -> Tuple: for config in sorted(self.test_config_path.glob('''**/*.yaml''' ) ): with self.subTest(config_file=__snake_case ): execute_subprocess_async( self.base_cmd + ['''--config_file''', str(__snake_case ), self.test_file_path] , env=os.environ.copy() ) def A ( self : str ) -> Optional[int]: execute_subprocess_async(['''accelerate''', '''test'''] , env=os.environ.copy() ) class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" lowerCamelCase__ = """test-tpu""" lowerCamelCase__ = """us-central1-a""" lowerCamelCase__ = """ls""" lowerCamelCase__ = ["""accelerate""", """tpu-config"""] lowerCamelCase__ = """cd /usr/share""" lowerCamelCase__ = """tests/test_samples/test_command_file.sh""" lowerCamelCase__ = """Running gcloud compute tpus tpu-vm ssh""" def A ( self : List[Any] ) -> Tuple: UpperCAmelCase : int = run_command( self.cmd + ['''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Tuple = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : str ) -> Optional[Any]: UpperCAmelCase : List[Any] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--debug'''] , return_stdout=__snake_case ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : str ) -> Optional[int]: UpperCAmelCase : Any = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : Optional[int] ) -> Tuple: UpperCAmelCase : Any = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--command''', '''echo "Hello World"''', '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase : List[Any] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command_file''', self.command_file, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : Tuple ) -> Any: UpperCAmelCase : int = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command_file''', self.command_file, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : int ) -> Any: UpperCAmelCase : Tuple = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> str: UpperCAmelCase : Optional[int] = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--accelerate_version''', '''12.0.0''', '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , )
23
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED UpperCamelCase_ = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } UpperCamelCase_ = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowerCamelCase_ ( ): '''simple docstring''' UpperCAmelCase_ : int = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) UpperCAmelCase_ : Dict = bs[:] UpperCAmelCase_ : Any = 0 for b in range(2**8 ): if b not in bs: bs.append(_a ) cs.append(2**8 + n ) n += 1 UpperCAmelCase_ : Any = [chr(_a ) for n in cs] return dict(zip(_a , _a ) ) def lowerCamelCase_ ( _a : List[str] ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = set() UpperCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase_ : Optional[int] = char return pairs class _snake_case ( __snake_case ): '''simple docstring''' A__ : str = VOCAB_FILES_NAMES A__ : List[str] = PRETRAINED_VOCAB_FILES_MAP A__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[int] = ["input_ids", "attention_mask"] def __init__( self: Union[str, Any] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Any ,lowerCamelCase_: Union[str, Any]="replace" ,lowerCamelCase_: Optional[Any]="<s>" ,lowerCamelCase_: List[Any]="</s>" ,lowerCamelCase_: List[str]="</s>" ,lowerCamelCase_: int="<s>" ,lowerCamelCase_: int="<unk>" ,lowerCamelCase_: str="<pad>" ,lowerCamelCase_: Optional[Any]="<mask>" ,lowerCamelCase_: List[str]=False ,**lowerCamelCase_: Tuple ,) -> Any: UpperCAmelCase_ : Union[str, Any] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else bos_token UpperCAmelCase_ : int = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else eos_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else sep_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else cls_token UpperCAmelCase_ : Optional[Any] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else unk_token UpperCAmelCase_ : List[str] = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ : str = AddedToken(lowerCamelCase_ ,lstrip=lowerCamelCase_ ,rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ ,lowerCamelCase_ ) else mask_token super().__init__( errors=lowerCamelCase_ ,bos_token=lowerCamelCase_ ,eos_token=lowerCamelCase_ ,unk_token=lowerCamelCase_ ,sep_token=lowerCamelCase_ ,cls_token=lowerCamelCase_ ,pad_token=lowerCamelCase_ ,mask_token=lowerCamelCase_ ,add_prefix_space=lowerCamelCase_ ,**lowerCamelCase_ ,) with open(lowerCamelCase_ ,encoding="""utf-8""" ) as vocab_handle: UpperCAmelCase_ : Union[str, Any] = json.load(lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} UpperCAmelCase_ : Any = errors # how to handle errors in decoding UpperCAmelCase_ : int = bytes_to_unicode() UpperCAmelCase_ : Dict = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase_ ,encoding="""utf-8""" ) as merges_handle: UpperCAmelCase_ : Any = merges_handle.read().split("""\n""" )[1:-1] UpperCAmelCase_ : int = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase_ : Union[str, Any] = dict(zip(lowerCamelCase_ ,range(len(lowerCamelCase_ ) ) ) ) UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : Optional[int] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase_ : int = re.compile(R"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def A__ ( self: List[str] ) -> List[str]: return len(self.encoder ) def A__ ( self: Any ) -> Union[str, Any]: return dict(self.encoder ,**self.added_tokens_encoder ) def A__ ( self: Tuple ,lowerCamelCase_: Dict ) -> Optional[Any]: if token in self.cache: return self.cache[token] UpperCAmelCase_ : Union[str, Any] = tuple(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = get_pairs(lowerCamelCase_ ) if not pairs: return token while True: UpperCAmelCase_ : Union[str, Any] = min(lowerCamelCase_ ,key=lambda lowerCamelCase_ : self.bpe_ranks.get(lowerCamelCase_ ,float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase_ , UpperCAmelCase_ : Any = bigram UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : List[str] = 0 while i < len(lowerCamelCase_ ): try: UpperCAmelCase_ : str = word.index(lowerCamelCase_ ,lowerCamelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase_ : Union[str, Any] = j if word[i] == first and i < len(lowerCamelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase_ : List[str] = tuple(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = new_word if len(lowerCamelCase_ ) == 1: break else: UpperCAmelCase_ : List[str] = get_pairs(lowerCamelCase_ ) UpperCAmelCase_ : int = """ """.join(lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = word return word def A__ ( self: Union[str, Any] ,lowerCamelCase_: Tuple ) -> List[str]: UpperCAmelCase_ : str = [] for token in re.findall(self.pat ,lowerCamelCase_ ): UpperCAmelCase_ : List[Any] = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase_ ).split(""" """ ) ) return bpe_tokens def A__ ( self: List[Any] ,lowerCamelCase_: Optional[Any] ) -> Optional[int]: return self.encoder.get(lowerCamelCase_ ,self.encoder.get(self.unk_token ) ) def A__ ( self: List[str] ,lowerCamelCase_: str ) -> Optional[Any]: return self.decoder.get(lowerCamelCase_ ) def A__ ( self: List[str] ,lowerCamelCase_: List[str] ) -> List[Any]: UpperCAmelCase_ : str = """""".join(lowerCamelCase_ ) UpperCAmelCase_ : int = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" ,errors=self.errors ) return text def A__ ( self: Optional[Any] ,lowerCamelCase_: str ,lowerCamelCase_: Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : List[Any] = os.path.join( lowerCamelCase_ ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase_ : List[str] = os.path.join( lowerCamelCase_ ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCamelCase_ ,"""w""" ,encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase_ ,ensure_ascii=lowerCamelCase_ ) + """\n""" ) UpperCAmelCase_ : str = 0 with open(lowerCamelCase_ ,"""w""" ,encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase_ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) UpperCAmelCase_ : Tuple = token_index writer.write(""" """.join(lowerCamelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file def A__ ( self: str ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ : int = [self.cls_token_id] UpperCAmelCase_ : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self: Union[str, Any] ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ,lowerCamelCase_: bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ ,token_ids_a=lowerCamelCase_ ,already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase_ )) + [1] return [1] + ([0] * len(lowerCamelCase_ )) + [1, 1] + ([0] * len(lowerCamelCase_ )) + [1] def A__ ( self: str ,lowerCamelCase_: List[int] ,lowerCamelCase_: Optional[List[int]] = None ) -> List[int]: UpperCAmelCase_ : Optional[Any] = [self.sep_token_id] UpperCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self: Optional[Any] ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: str=False ,**lowerCamelCase_: List[str] ) -> Optional[int]: UpperCAmelCase_ : Optional[int] = kwargs.pop("""add_prefix_space""" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase_ ) > 0 and not text[0].isspace()): UpperCAmelCase_ : Dict = """ """ + text return (text, kwargs) def A__ ( self: List[str] ,lowerCamelCase_: Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCamelCase_: Optional[int] = None ,lowerCamelCase_: PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCamelCase_: Optional[int] = None ,lowerCamelCase_: Optional[bool] = None ,) -> dict: UpperCAmelCase_ : Optional[int] = super()._pad( encoded_inputs=lowerCamelCase_ ,max_length=lowerCamelCase_ ,padding_strategy=lowerCamelCase_ ,pad_to_multiple_of=lowerCamelCase_ ,return_attention_mask=lowerCamelCase_ ,) # Load from model defaults if return_attention_mask is None: UpperCAmelCase_ : str = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: UpperCAmelCase_ : str = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. UpperCAmelCase_ : List[Any] = len(encoded_inputs["""global_attention_mask"""] ) != len(lowerCamelCase_ ) if needs_to_be_padded: UpperCAmelCase_ : Dict = len(lowerCamelCase_ ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` UpperCAmelCase_ : str = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": UpperCAmelCase_ : List[str] = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
345
0
"""simple docstring""" import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : Dict ="facebook/bart-large-mnli" a : Optional[int] =( "This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which " "should be the text to classify, and `labels`, which should be the list of labels to use for classification. " "It returns the most likely label in the list of provided `labels` for the input text." ) a : str ="text_classifier" a : Dict =AutoTokenizer a : List[str] =AutoModelForSequenceClassification a : List[str] =["text", ["text"]] a : str =["text"] def lowercase__ ( self ): """simple docstring""" super().setup() lowerCAmelCase : Any = self.model.config lowerCAmelCase : Optional[Any] = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("entail" ): lowerCAmelCase : List[str] = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("Could not determine the entailment ID from the model config, please pass it at init." ) def lowercase__ ( self , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = labels return self.pre_processor( [text] * len(snake_case__ ) , [f"""This example is {label}""" for label in labels] , return_tensors="pt" , padding="max_length" , ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : Any = outputs.logits lowerCAmelCase : List[str] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
353
"""simple docstring""" import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , **snake_case__ , ): """simple docstring""" super().__init__(features=snake_case__ , cache_dir=snake_case__ , keep_in_memory=snake_case__ , **snake_case__ ) lowerCAmelCase : Dict = Sql( cache_dir=snake_case__ , features=snake_case__ , sql=snake_case__ , con=snake_case__ , **snake_case__ , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Tuple = None lowerCAmelCase : Optional[int] = None lowerCAmelCase : str = None lowerCAmelCase : int = None self.builder.download_and_prepare( download_config=snake_case__ , download_mode=snake_case__ , verification_mode=snake_case__ , base_path=snake_case__ , ) # Build dataset for splits lowerCAmelCase : str = self.builder.as_dataset( split="train" , verification_mode=snake_case__ , in_memory=self.keep_in_memory ) return dataset class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None , **snake_case__ , ): """simple docstring""" if num_proc is not None and num_proc <= 0: raise ValueError(f"""num_proc {num_proc} must be an integer > 0.""" ) lowerCAmelCase : Tuple = dataset lowerCAmelCase : Tuple = name lowerCAmelCase : List[str] = con lowerCAmelCase : List[Any] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE lowerCAmelCase : Optional[int] = num_proc lowerCAmelCase : Optional[int] = to_sql_kwargs def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[str] = self.to_sql_kwargs.pop("sql" , snake_case__ ) lowerCAmelCase : List[Any] = self.to_sql_kwargs.pop("con" , snake_case__ ) lowerCAmelCase : Dict = self.to_sql_kwargs.pop("index" , snake_case__ ) lowerCAmelCase : Dict = self._write(index=snake_case__ , **self.to_sql_kwargs ) return written def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase , lowerCAmelCase , lowerCAmelCase : Optional[int] = args lowerCAmelCase : Optional[int] = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs lowerCAmelCase : List[Any] = query_table( table=self.dataset.data , key=slice(snake_case__ , offset + self.batch_size ) , indices=self.dataset._indices , ) lowerCAmelCase : Tuple = batch.to_pandas() lowerCAmelCase : str = df.to_sql(self.name , self.con , index=snake_case__ , **snake_case__ ) return num_rows or len(snake_case__ ) def lowercase__ ( self , snake_case__ , **snake_case__ ): """simple docstring""" lowerCAmelCase : Tuple = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: lowerCAmelCase , lowerCAmelCase : Dict = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql , [(offset, index, to_sql_kwargs) for offset in range(0 , snake_case__ , snake_case__ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += num_rows return written
133
0
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class lowerCamelCase (_snake_case ): '''simple docstring''' def __UpperCAmelCase ( self , _UpperCamelCase ) -> float: return 0.0 def lowercase__ ( __snake_case : np.ndarray , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : int = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) UpperCAmelCase_ : Dict = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def lowercase__ ( __snake_case : FilterType , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : List[str] = 512 UpperCAmelCase_ : str = [1] + [0] * (size - 1) UpperCAmelCase_ : Optional[Any] = [filter_type.process(__snake_case ) for item in inputs] UpperCAmelCase_ : Dict = [0] * (samplerate - size) # zero-padding outputs += filler UpperCAmelCase_ : Optional[int] = np.abs(np.fft.fft(__snake_case ) ) UpperCAmelCase_ : List[str] = 20 * np.logaa(__snake_case ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds UpperCAmelCase_ : Union[str, Any] = get_bounds(__snake_case , __snake_case ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__snake_case ) plt.show() def lowercase__ ( __snake_case : FilterType , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : int = 512 UpperCAmelCase_ : Tuple = [1] + [0] * (size - 1) UpperCAmelCase_ : Tuple = [filter_type.process(__snake_case ) for item in inputs] UpperCAmelCase_ : List[str] = [0] * (samplerate - size) # zero-padding outputs += filler UpperCAmelCase_ : Dict = np.angle(np.fft.fft(__snake_case ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__snake_case , -2 * pi ) ) plt.show()
29
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np UpperCAmelCase : Union[str, Any] = re.compile(r"\b(a|an|the)\b", re.UNICODE) UpperCAmelCase : Optional[Any] = None def _SCREAMING_SNAKE_CASE () -> List[Any]: '''simple docstring''' lowercase_ = argparse.ArgumentParser("""Official evaluation script for SQuAD version 2.0.""" ) parser.add_argument("""data_file""" , metavar="""data.json""" , help="""Input data JSON file.""" ) parser.add_argument("""pred_file""" , metavar="""pred.json""" , help="""Model predictions.""" ) parser.add_argument( """--out-file""" , """-o""" , metavar="""eval.json""" , help="""Write accuracy metrics to file (default is stdout).""" ) parser.add_argument( """--na-prob-file""" , """-n""" , metavar="""na_prob.json""" , help="""Model estimates of probability of no answer.""" ) parser.add_argument( """--na-prob-thresh""" , """-t""" , type=__lowerCAmelCase , default=1.0 , help="""Predict \"\" if no-answer probability exceeds this (default = 1.0).""" , ) parser.add_argument( """--out-image-dir""" , """-p""" , metavar="""out_images""" , default=__lowerCAmelCase , help="""Save precision-recall curves to directory.""" ) parser.add_argument("""--verbose""" , """-v""" , action="""store_true""" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase_ = bool(qa["""answers"""]["""text"""] ) return qid_to_has_ans def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Optional[int]: '''simple docstring''' def remove_articles(__lowerCAmelCase ): return ARTICLES_REGEX.sub(""" """ , __lowerCAmelCase ) def white_space_fix(__lowerCAmelCase ): return " ".join(text.split() ) def remove_punc(__lowerCAmelCase ): lowercase_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__lowerCAmelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__lowerCAmelCase ) ) ) ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> List[Any]: '''simple docstring''' if not s: return [] return normalize_answer(__lowerCAmelCase ).split() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: '''simple docstring''' return int(normalize_answer(__lowerCAmelCase ) == normalize_answer(__lowerCAmelCase ) ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> List[str]: '''simple docstring''' lowercase_ = get_tokens(__lowerCAmelCase ) lowercase_ = get_tokens(__lowerCAmelCase ) lowercase_ = collections.Counter(__lowerCAmelCase ) & collections.Counter(__lowerCAmelCase ) lowercase_ = sum(common.values() ) if len(__lowerCAmelCase ) == 0 or len(__lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase_ = 1.0 * num_same / len(__lowerCAmelCase ) lowercase_ = 1.0 * num_same / len(__lowerCAmelCase ) lowercase_ = (2 * precision * recall) / (precision + recall) return fa def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' lowercase_ = {} lowercase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase_ = qa["""id"""] lowercase_ = [t for t in qa["""answers"""]["""text"""] if normalize_answer(__lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase_ = [""""""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue lowercase_ = preds[qid] # Take max over all gold answers lowercase_ = max(compute_exact(__lowerCAmelCase , __lowerCAmelCase ) for a in gold_answers ) lowercase_ = max(compute_fa(__lowerCAmelCase , __lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' lowercase_ = {} for qid, s in scores.items(): lowercase_ = na_probs[qid] > na_prob_thresh if pred_na: lowercase_ = float(not qid_to_has_ans[qid] ) else: lowercase_ = s return new_scores def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None ) -> List[str]: '''simple docstring''' if not qid_list: lowercase_ = len(__lowerCAmelCase ) return collections.OrderedDict( [ ("""exact""", 100.0 * sum(exact_scores.values() ) / total), ("""f1""", 100.0 * sum(fa_scores.values() ) / total), ("""total""", total), ] ) else: lowercase_ = len(__lowerCAmelCase ) return collections.OrderedDict( [ ("""exact""", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("""f1""", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("""total""", total), ] ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' for k in new_eval: lowercase_ = new_eval[k] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[str]: '''simple docstring''' plt.step(__lowerCAmelCase , __lowerCAmelCase , color="""b""" , alpha=0.2 , where="""post""" ) plt.fill_between(__lowerCAmelCase , __lowerCAmelCase , step="""post""" , alpha=0.2 , color="""b""" ) plt.xlabel("""Recall""" ) plt.ylabel("""Precision""" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(__lowerCAmelCase ) plt.savefig(__lowerCAmelCase ) plt.clf() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None ) -> List[Any]: '''simple docstring''' lowercase_ = sorted(__lowerCAmelCase , key=lambda __lowerCAmelCase : na_probs[k] ) lowercase_ = 0.0 lowercase_ = 1.0 lowercase_ = 0.0 lowercase_ = [1.0] lowercase_ = [0.0] lowercase_ = 0.0 for i, qid in enumerate(__lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase_ = true_pos / float(i + 1 ) lowercase_ = true_pos / float(__lowerCAmelCase ) if i == len(__lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(__lowerCAmelCase ) recalls.append(__lowerCAmelCase ) if out_image: plot_pr_curve(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) return {"ap": 100.0 * avg_prec} def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Dict: '''simple docstring''' if out_image_dir and not os.path.exists(__lowerCAmelCase ): os.makedirs(__lowerCAmelCase ) lowercase_ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase_ = make_precision_recall_eval( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , out_image=os.path.join(__lowerCAmelCase , """pr_exact.png""" ) , title="""Precision-Recall curve for Exact Match score""" , ) lowercase_ = make_precision_recall_eval( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , out_image=os.path.join(__lowerCAmelCase , """pr_f1.png""" ) , title="""Precision-Recall curve for F1 score""" , ) lowercase_ = {k: float(__lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase_ = make_precision_recall_eval( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , out_image=os.path.join(__lowerCAmelCase , """pr_oracle.png""" ) , title="""Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)""" , ) merge_eval(__lowerCAmelCase , __lowerCAmelCase , """pr_exact""" ) merge_eval(__lowerCAmelCase , __lowerCAmelCase , """pr_f1""" ) merge_eval(__lowerCAmelCase , __lowerCAmelCase , """pr_oracle""" ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[str]: '''simple docstring''' if not qid_list: return lowercase_ = [na_probs[k] for k in qid_list] lowercase_ = np.ones_like(__lowerCAmelCase ) / float(len(__lowerCAmelCase ) ) plt.hist(__lowerCAmelCase , weights=__lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel("""Model probability of no-answer""" ) plt.ylabel("""Proportion of dataset""" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(__lowerCAmelCase , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase_ = num_no_ans lowercase_ = cur_score lowercase_ = 0.0 lowercase_ = sorted(__lowerCAmelCase , key=lambda __lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(__lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase_ = scores[qid] else: if preds[qid]: lowercase_ = -1 else: lowercase_ = 0 cur_score += diff if cur_score > best_score: lowercase_ = cur_score lowercase_ = na_probs[qid] return 100.0 * best_score / len(__lowerCAmelCase ), best_thresh def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: '''simple docstring''' lowercase_ , lowercase_ = find_best_thresh(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) lowercase_ , lowercase_ = find_best_thresh(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) lowercase_ = best_exact lowercase_ = exact_thresh lowercase_ = best_fa lowercase_ = fa_thresh def _SCREAMING_SNAKE_CASE () -> int: '''simple docstring''' with open(OPTS.data_file ) as f: lowercase_ = json.load(__lowerCAmelCase ) lowercase_ = dataset_json["""data"""] with open(OPTS.pred_file ) as f: lowercase_ = json.load(__lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase_ = json.load(__lowerCAmelCase ) else: lowercase_ = {k: 0.0 for k in preds} lowercase_ = make_qid_to_has_ans(__lowerCAmelCase ) # maps qid to True/False lowercase_ = [k for k, v in qid_to_has_ans.items() if v] lowercase_ = [k for k, v in qid_to_has_ans.items() if not v] lowercase_ , lowercase_ = get_raw_scores(__lowerCAmelCase , __lowerCAmelCase ) lowercase_ = apply_no_ans_threshold(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , OPTS.na_prob_thresh ) lowercase_ = apply_no_ans_threshold(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , OPTS.na_prob_thresh ) lowercase_ = make_eval_dict(__lowerCAmelCase , __lowerCAmelCase ) if has_ans_qids: lowercase_ = make_eval_dict(__lowerCAmelCase , __lowerCAmelCase , qid_list=__lowerCAmelCase ) merge_eval(__lowerCAmelCase , __lowerCAmelCase , """HasAns""" ) if no_ans_qids: lowercase_ = make_eval_dict(__lowerCAmelCase , __lowerCAmelCase , qid_list=__lowerCAmelCase ) merge_eval(__lowerCAmelCase , __lowerCAmelCase , """NoAns""" ) if OPTS.na_prob_file: find_all_best_thresh(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(__lowerCAmelCase , __lowerCAmelCase , OPTS.out_image_dir , """hasAns""" ) histogram_na_prob(__lowerCAmelCase , __lowerCAmelCase , OPTS.out_image_dir , """noAns""" ) if OPTS.out_file: with open(OPTS.out_file , """w""" ) as f: json.dump(__lowerCAmelCase , __lowerCAmelCase ) else: print(json.dumps(__lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": UpperCAmelCase : Union[str, Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
136
0
'''simple docstring''' def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _UpperCAmelCase : Any = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(lowerCAmelCase_ ) ) return round(lowerCAmelCase_ , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
351
'''simple docstring''' from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal lowerCAmelCase_ : str = logging.get_logger(__name__) lowerCAmelCase_ : Union[str, Any] = TypeVar('''DatasetType''', Dataset, IterableDataset) def __A ( lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = "first_exhausted" , ): from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError("""Unable to interleave an empty list of datasets.""" ) for i, dataset in enumerate(lowerCAmelCase_ ): if not isinstance(lowerCAmelCase_ , (Dataset, IterableDataset) ): if isinstance(lowerCAmelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " """is an empty dataset dictionary.""" ) raise ValueError( f"Dataset at position {i} has at least one split: {list(lowerCAmelCase_ )}\n" f"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(lowerCAmelCase_ ) )}']" ) raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowerCAmelCase_ ).__name__}." ) if i == 0: _UpperCAmelCase , _UpperCAmelCase : Dict = ( (Dataset, IterableDataset) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError( f"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(f"{stopping_strategy} is not supported. Please enter a valid stopping_strategy." ) if dataset_type is Dataset: return _interleave_map_style_datasets( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , info=lowerCAmelCase_ , split=lowerCAmelCase_ , stopping_strategy=lowerCAmelCase_ ) else: return _interleave_iterable_datasets( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , info=lowerCAmelCase_ , split=lowerCAmelCase_ , stopping_strategy=lowerCAmelCase_ ) def __A ( lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = 0 , ): if not dsets: raise ValueError("""Unable to concatenate an empty list of datasets.""" ) for i, dataset in enumerate(lowerCAmelCase_ ): if not isinstance(lowerCAmelCase_ , (Dataset, IterableDataset) ): if isinstance(lowerCAmelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " """is an empty dataset dictionary.""" ) raise ValueError( f"Dataset at position {i} has at least one split: {list(lowerCAmelCase_ )}\n" f"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(lowerCAmelCase_ ) )}']" ) raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowerCAmelCase_ ).__name__}." ) if i == 0: _UpperCAmelCase , _UpperCAmelCase : Dict = ( (Dataset, IterableDataset) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError( f"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if dataset_type is Dataset: return _concatenate_map_style_datasets(lowerCAmelCase_ , info=lowerCAmelCase_ , split=lowerCAmelCase_ , axis=lowerCAmelCase_ ) else: return _concatenate_iterable_datasets(lowerCAmelCase_ , info=lowerCAmelCase_ , split=lowerCAmelCase_ , axis=lowerCAmelCase_ )
170
0
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _SCREAMING_SNAKE_CASE : Dict = get_logger(__name__) class A__ : """simple docstring""" def __init__( self , __snake_case = None ): snake_case = ( os.path.join(__snake_case , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) snake_case = Extractor def a_ ( self , __snake_case ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" snake_case = os.path.abspath(__snake_case ) return os.path.join(self.extract_dir , hash_url_to_filename(__snake_case ) ) def a_ ( self , __snake_case , __snake_case ): return force_extract or ( not os.path.isfile(__snake_case ) and not (os.path.isdir(__snake_case ) and os.listdir(__snake_case )) ) def a_ ( self , __snake_case , __snake_case = False ): snake_case = self.extractor.infer_extractor_format(__snake_case ) if not extractor_format: return input_path snake_case = self._get_output_path(__snake_case ) if self._do_extract(__snake_case , __snake_case ): self.extractor.extract(__snake_case , __snake_case , __snake_case ) return output_path class A__ ( snake_case__ ): """simple docstring""" @classmethod @abstractmethod def a_ ( cls , __snake_case , **__snake_case ): ... @staticmethod @abstractmethod def a_ ( __snake_case , __snake_case ): ... class A__ ( snake_case__ , snake_case__ ): """simple docstring""" __magic_name__ = [] @staticmethod def a_ ( __snake_case , __snake_case ): with open(__snake_case , '''rb''' ) as f: return f.read(__snake_case ) @classmethod def a_ ( cls , __snake_case , __snake_case = b"" ): if not magic_number: snake_case = max(len(__snake_case ) for cls_magic_number in cls.magic_numbers ) try: snake_case = cls.read_magic_number(__snake_case , __snake_case ) except OSError: return False return any(magic_number.startswith(__snake_case ) for cls_magic_number in cls.magic_numbers ) class A__ ( snake_case__ ): """simple docstring""" @classmethod def a_ ( cls , __snake_case , **__snake_case ): return tarfile.is_tarfile(__snake_case ) @staticmethod def a_ ( __snake_case , __snake_case ): def resolved(__snake_case ) -> str: return os.path.realpath(os.path.abspath(__snake_case ) ) def badpath(__snake_case , __snake_case ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__snake_case , __snake_case ) ).startswith(__snake_case ) def badlink(__snake_case , __snake_case ) -> bool: # Links are interpreted relative to the directory containing the link snake_case = resolved(os.path.join(__snake_case , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__snake_case ) snake_case = resolved(__snake_case ) for finfo in members: if badpath(finfo.name , __snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__snake_case , __snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__snake_case , __snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def a_ ( __snake_case , __snake_case ): os.makedirs(__snake_case , exist_ok=__snake_case ) snake_case = tarfile.open(__snake_case ) tar_file.extractall(__snake_case , members=TarExtractor.safemembers(__snake_case , __snake_case ) ) tar_file.close() class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\x1F\x8B'] @staticmethod def a_ ( __snake_case , __snake_case ): with gzip.open(__snake_case , '''rb''' ) as gzip_file: with open(__snake_case , '''wb''' ) as extracted_file: shutil.copyfileobj(__snake_case , __snake_case ) class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [ b'PK\x03\x04', b'PK\x05\x06', # empty archive b'PK\x07\x08', # spanned archive ] @classmethod def a_ ( cls , __snake_case , __snake_case = b"" ): if super().is_extractable(__snake_case , magic_number=__snake_case ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__snake_case , '''rb''' ) as fp: snake_case = _EndRecData(__snake_case ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: snake_case = fp.read(__snake_case ) # CD is where we expect it to be if len(__snake_case ) == sizeCentralDir: snake_case = struct.unpack(__snake_case , __snake_case ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def a_ ( __snake_case , __snake_case ): os.makedirs(__snake_case , exist_ok=__snake_case ) with zipfile.ZipFile(__snake_case , '''r''' ) as zip_file: zip_file.extractall(__snake_case ) zip_file.close() class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\xFD\x37\x7A\x58\x5A\x00'] @staticmethod def a_ ( __snake_case , __snake_case ): with lzma.open(__snake_case ) as compressed_file: with open(__snake_case , '''wb''' ) as extracted_file: shutil.copyfileobj(__snake_case , __snake_case ) class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'Rar!\x1a\x07\x00', b'Rar!\x1a\x07\x01\x00'] # RAR_ID # RAR5_ID @staticmethod def a_ ( __snake_case , __snake_case ): if not config.RARFILE_AVAILABLE: raise ImportError('''Please pip install rarfile''' ) import rarfile os.makedirs(__snake_case , exist_ok=__snake_case ) snake_case = rarfile.RarFile(__snake_case ) rf.extractall(__snake_case ) rf.close() class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\x28\xb5\x2F\xFD'] @staticmethod def a_ ( __snake_case , __snake_case ): if not config.ZSTANDARD_AVAILABLE: raise ImportError('''Please pip install zstandard''' ) import zstandard as zstd snake_case = zstd.ZstdDecompressor() with open(__snake_case , '''rb''' ) as ifh, open(__snake_case , '''wb''' ) as ofh: dctx.copy_stream(__snake_case , __snake_case ) class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\x42\x5A\x68'] @staticmethod def a_ ( __snake_case , __snake_case ): with bza.open(__snake_case , '''rb''' ) as compressed_file: with open(__snake_case , '''wb''' ) as extracted_file: shutil.copyfileobj(__snake_case , __snake_case ) class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\x37\x7A\xBC\xAF\x27\x1C'] @staticmethod def a_ ( __snake_case , __snake_case ): if not config.PY7ZR_AVAILABLE: raise ImportError('''Please pip install py7zr''' ) import pyazr os.makedirs(__snake_case , exist_ok=__snake_case ) with pyazr.SevenZipFile(__snake_case , '''r''' ) as archive: archive.extractall(__snake_case ) class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [b'\x04\x22\x4D\x18'] @staticmethod def a_ ( __snake_case , __snake_case ): if not config.LZ4_AVAILABLE: raise ImportError('''Please pip install lz4''' ) import lza.frame with lza.frame.open(__snake_case , '''rb''' ) as compressed_file: with open(__snake_case , '''wb''' ) as extracted_file: shutil.copyfileobj(__snake_case , __snake_case ) class A__ : """simple docstring""" __magic_name__ = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def a_ ( cls ): return max( len(__snake_case ) for extractor in cls.extractors.values() if issubclass(__snake_case , __snake_case ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def a_ ( __snake_case , __snake_case ): try: return MagicNumberBaseExtractor.read_magic_number(__snake_case , magic_number_length=__snake_case ) except OSError: return b"" @classmethod def a_ ( cls , __snake_case , __snake_case = False ): warnings.warn( '''Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'infer_extractor_format\' instead.''' , category=__snake_case , ) snake_case = cls.infer_extractor_format(__snake_case ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def a_ ( cls , __snake_case ): # <Added version="2.4.0"/> snake_case = cls._get_magic_number_max_length() snake_case = cls._read_magic_number(__snake_case , __snake_case ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__snake_case , magic_number=__snake_case ): return extractor_format @classmethod def a_ ( cls , __snake_case , __snake_case , __snake_case = None , __snake_case = "deprecated" , ): os.makedirs(os.path.dirname(__snake_case ) , exist_ok=__snake_case ) # Prevent parallel extractions snake_case = str(Path(__snake_case ).with_suffix('''.lock''' ) ) with FileLock(__snake_case ): shutil.rmtree(__snake_case , ignore_errors=__snake_case ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__snake_case , __snake_case ): # passed as positional arg warnings.warn( '''Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'extractor_format\' instead.''' , category=__snake_case , ) snake_case = extractor if extractor != '''deprecated''' else extractor_format else: snake_case = cls.extractors[extractor_format] return extractor.extract(__snake_case , __snake_case ) else: warnings.warn( '''Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an ''' '''exception in 3.0.0.''' , category=__snake_case , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__snake_case ): return extractor.extract(__snake_case , __snake_case )
127
import random def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" snake_case = a[left_index] snake_case = left_index + 1 for j in range(left_index + 1 ,UpperCamelCase_ ): if a[j] < pivot: snake_case , snake_case = a[i], a[j] i += 1 snake_case , snake_case = a[i - 1], a[left_index] return i - 1 def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" if left < right: snake_case = random.randint(UpperCamelCase_ ,right - 1 ) snake_case , snake_case = ( a[left], a[pivot], ) # switches the pivot with the left most bound snake_case = partition(UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ) quick_sort_random( UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ) # recursive quicksort to the left of the pivot point quick_sort_random( UpperCamelCase_ ,pivot_index + 1 ,UpperCamelCase_ ) # recursive quicksort to the right of the pivot point def UpperCAmelCase__ (): """simple docstring""" snake_case = input('''Enter numbers separated by a comma:\n''' ).strip() snake_case = [int(UpperCamelCase_ ) for item in user_input.split(''',''' )] quick_sort_random(UpperCamelCase_ ,0 ,len(UpperCamelCase_ ) ) print(UpperCamelCase_ ) if __name__ == "__main__": main()
127
1
'''simple docstring''' def _snake_case ( A ) -> Any: if any(not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or x < 0 for x in sequence ): raise TypeError('''Sequence must be list of non-negative integers''' ) for _ in range(len(_lowerCAmelCase ) ): for i, (rod_upper, rod_lower) in enumerate(zip(_lowerCAmelCase , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
354
'''simple docstring''' import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = '''https://openaipublic.azureedge.net/jukebox/models/''' __UpperCAmelCase = { '''jukebox-1b-lyrics''': [ '''5b/vqvae.pth.tar''', '''5b/prior_level_0.pth.tar''', '''5b/prior_level_1.pth.tar''', '''1b_lyrics/prior_level_2.pth.tar''', ], '''jukebox-5b-lyrics''': [ '''5b/vqvae.pth.tar''', '''5b/prior_level_0.pth.tar''', '''5b/prior_level_1.pth.tar''', '''5b_lyrics/prior_level_2.pth.tar''', ], } def _snake_case ( A ) -> Union[str, Any]: if key.endswith('''.model.1.bias''' ) and len(key.split('''.''' ) ) > 10: lowerCAmelCase__ = key.replace('''.model.1.bias''' , '''.conv1d_1.bias''' ) elif key.endswith('''.model.1.weight''' ) and len(key.split('''.''' ) ) > 10: lowerCAmelCase__ = key.replace('''.model.1.weight''' , '''.conv1d_1.weight''' ) elif key.endswith('''.model.3.bias''' ) and len(key.split('''.''' ) ) > 10: lowerCAmelCase__ = key.replace('''.model.3.bias''' , '''.conv1d_2.bias''' ) elif key.endswith('''.model.3.weight''' ) and len(key.split('''.''' ) ) > 10: lowerCAmelCase__ = key.replace('''.model.3.weight''' , '''.conv1d_2.weight''' ) if "conditioner_blocks.0." in key: lowerCAmelCase__ = key.replace('''conditioner_blocks.0''' , '''conditioner_blocks''' ) if "prime_prior" in key: lowerCAmelCase__ = key.replace('''prime_prior''' , '''encoder''' ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: lowerCAmelCase__ = key.replace('''.emb.''' , '''.''' ) if key.endswith('''k''' ): # replace vqvae.X.k with vqvae.X.codebook return key.replace('''.k''' , '''.codebook''' ) if "y_emb." in key: return key.replace('''y_emb.''' , '''metadata_embedding.''' ) if "x_emb.emb." in key: lowerCAmelCase__ = key.replace('''0.x_emb.emb''' , '''embed_tokens''' ) if "prime_state_ln" in key: return key.replace('''prime_state_ln''' , '''encoder.final_layer_norm''' ) if ".ln" in key: return key.replace('''.ln''' , '''.layer_norm''' ) if "_ln" in key: return key.replace('''_ln''' , '''_layer_norm''' ) if "prime_state_proj" in key: return key.replace('''prime_state_proj''' , '''encoder.proj_in''' ) if "prime_x_out" in key: return key.replace('''prime_x_out''' , '''encoder.lm_head''' ) if "prior.x_out" in key: return key.replace('''x_out''' , '''fc_proj_out''' ) if "x_emb" in key: return key.replace('''x_emb''' , '''embed_tokens''' ) return key def _snake_case ( A , A , A , A ) -> Optional[int]: lowerCAmelCase__ = {} import re lowerCAmelCase__ = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) lowerCAmelCase__ = re.compile( R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) lowerCAmelCase__ = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) lowerCAmelCase__ = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) lowerCAmelCase__ = re.compile( R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) lowerCAmelCase__ = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) lowerCAmelCase__ = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)''' ) lowerCAmelCase__ = re.compile( R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) lowerCAmelCase__ = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)''' ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(A ): lowerCAmelCase__ = re_encoder_block_conv_in.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) lowerCAmelCase__ = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}""" lowerCAmelCase__ = re_encoder_block_conv_in.sub(A , A ) elif re_encoder_block_resnet.fullmatch(A ): lowerCAmelCase__ = re_encoder_block_resnet.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) lowerCAmelCase__ = {'''1''': 1, '''3''': 2}[groups[-2]] lowerCAmelCase__ = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.""" lowerCAmelCase__ = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" lowerCAmelCase__ = prefix + resnet_block lowerCAmelCase__ = re_encoder_block_resnet.sub(A , A ) elif re_encoder_block_proj_out.fullmatch(A ): lowerCAmelCase__ = re_encoder_block_proj_out.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}""" lowerCAmelCase__ = re_encoder_block_proj_out.sub(A , A ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(A ): lowerCAmelCase__ = re_decoder_block_conv_out.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) - 2 lowerCAmelCase__ = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}""" lowerCAmelCase__ = re_decoder_block_conv_out.sub(A , A ) elif re_decoder_block_resnet.fullmatch(A ): lowerCAmelCase__ = re_decoder_block_resnet.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) - 2 lowerCAmelCase__ = {'''1''': 1, '''3''': 2}[groups[-2]] lowerCAmelCase__ = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.""" lowerCAmelCase__ = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" lowerCAmelCase__ = prefix + resnet_block lowerCAmelCase__ = re_decoder_block_resnet.sub(A , A ) elif re_decoder_block_proj_in.fullmatch(A ): lowerCAmelCase__ = re_decoder_block_proj_in.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}""" lowerCAmelCase__ = re_decoder_block_proj_in.sub(A , A ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(A ): lowerCAmelCase__ = re_prior_cond_conv_out.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[1] ) * 2 + int(groups[2] ) - 2 lowerCAmelCase__ = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}""" lowerCAmelCase__ = re_prior_cond_conv_out.sub(A , A ) elif re_prior_cond_resnet.fullmatch(A ): lowerCAmelCase__ = re_prior_cond_resnet.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = int(groups[1] ) * 2 + int(groups[2] ) - 2 lowerCAmelCase__ = {'''1''': 1, '''3''': 2}[groups[-2]] lowerCAmelCase__ = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.""" lowerCAmelCase__ = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" lowerCAmelCase__ = prefix + resnet_block lowerCAmelCase__ = re_prior_cond_resnet.sub(A , A ) elif re_prior_cond_proj_in.fullmatch(A ): lowerCAmelCase__ = re_prior_cond_proj_in.match(A ) lowerCAmelCase__ = regex_match.groups() lowerCAmelCase__ = F"""conditioner_blocks.upsampler.proj_in.{groups[-1]}""" lowerCAmelCase__ = re_prior_cond_proj_in.sub(A , A ) # keep original key else: lowerCAmelCase__ = original_key lowerCAmelCase__ = replace_key(A ) if F"""{key_prefix}.{key}""" not in model_state_dict or key is None: print(F"""failed converting {original_key} to {key}, does not match""" ) # handle missmatched shape elif value.shape != model_state_dict[F"""{key_prefix}.{key}"""].shape: lowerCAmelCase__ = model_state_dict[F"""{key_prefix}.{key}"""] print(F"""{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match""" ) lowerCAmelCase__ = original_key lowerCAmelCase__ = original_key lowerCAmelCase__ = value return new_dict @torch.no_grad() def _snake_case ( A=None , A=None ) -> str: for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""" ): lowerCAmelCase__ = requests.get(F"""{PREFIX}{file}""" , allow_redirects=A ) os.makedirs(F"""{pytorch_dump_folder_path}/""" , exist_ok=A ) open(F"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""" , '''wb''' ).write(r.content ) lowerCAmelCase__ = MODEL_MAPPING[model_name.split('''/''' )[-1]] lowerCAmelCase__ = JukeboxConfig.from_pretrained(A ) lowerCAmelCase__ = JukeboxModel(A ) lowerCAmelCase__ = [] lowerCAmelCase__ = {} for i, dict_name in enumerate(A ): lowerCAmelCase__ = torch.load(F"""{pytorch_dump_folder_path}/{dict_name.split("/" )[-1]}""" )['''model'''] lowerCAmelCase__ = {} for k in old_dic.keys(): if k.endswith('''.b''' ): lowerCAmelCase__ = old_dic[k] elif k.endswith('''.w''' ): lowerCAmelCase__ = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: lowerCAmelCase__ = old_dic[k] else: lowerCAmelCase__ = old_dic[k] lowerCAmelCase__ = '''vqvae''' if i == 0 else F"""priors.{3 - i}""" lowerCAmelCase__ = fix_jukebox_keys(A , model.state_dict() , A , A ) weight_dict.append(A ) lowerCAmelCase__ = weight_dict.pop(0 ) model.vqvae.load_state_dict(A ) for i in range(len(A ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(A ).mkdir(exist_ok=A ) with open(F"""{pytorch_dump_folder_path}/mapping.json""" , '''w''' ) as txtfile: json.dump(A , A ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(A ) return weight_dict if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''jukebox-5b-lyrics''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''jukebox-5b-lyrics-converted''', type=str, help='''Path to the output PyTorch model directory.''', ) __UpperCAmelCase = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
228
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class __lowerCAmelCase ( a ): """simple docstring""" def __init__( self : Tuple , _lowerCAmelCase : Union[str, Any] ) -> List[Any]: """simple docstring""" snake_case_ = data def __iter__( self : Optional[int] ) -> Optional[int]: """simple docstring""" for element in self.data: yield element def _lowerCAmelCase ( lowerCAmelCase_ :List[Any]=True )->Optional[Any]: '''simple docstring''' snake_case_ = Accelerator(even_batches=lowerCAmelCase_ ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def _lowerCAmelCase ( lowerCAmelCase_ :Accelerator , lowerCAmelCase_ :int , lowerCAmelCase_ :int , lowerCAmelCase_ :bool = False )->str: '''simple docstring''' if iterable: snake_case_ = DummyIterableDataset(torch.as_tensor(range(lowerCAmelCase_ ) ) ) else: snake_case_ = TensorDataset(torch.as_tensor(range(lowerCAmelCase_ ) ) ) snake_case_ = DataLoader(lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) snake_case_ = accelerator.prepare(lowerCAmelCase_ ) return dl def _lowerCAmelCase ( lowerCAmelCase_ :Accelerator , lowerCAmelCase_ :int , lowerCAmelCase_ :int , lowerCAmelCase_ :List[int] , lowerCAmelCase_ :List[int] , )->str: '''simple docstring''' snake_case_ = create_dataloader(accelerator=lowerCAmelCase_ , dataset_size=lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) snake_case_ = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def _lowerCAmelCase ( )->str: '''simple docstring''' snake_case_ = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def _lowerCAmelCase ( )->int: '''simple docstring''' snake_case_ = create_accelerator(even_batches=lowerCAmelCase_ ) verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def _lowerCAmelCase ( )->Optional[Any]: '''simple docstring''' snake_case_ = create_accelerator(even_batches=lowerCAmelCase_ ) snake_case_ = torch.nn.Linear(1 , 1 ) snake_case_ = accelerator.prepare(lowerCAmelCase_ ) snake_case_ = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) snake_case_ = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(lowerCAmelCase_ ): snake_case_ = ddp_model(batch[0].float() ) snake_case_ = output.sum() loss.backward() batch_idxs.append(lowerCAmelCase_ ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def _lowerCAmelCase ( lowerCAmelCase_ :Tuple )->int: '''simple docstring''' with warnings.catch_warnings(record=lowerCAmelCase_ ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , lowerCAmelCase_ ) assert "only supported for multi-GPU" in str(w[-1].message ) def _lowerCAmelCase ( )->str: '''simple docstring''' snake_case_ = True snake_case_ = False snake_case_ = create_accelerator(even_batches=lowerCAmelCase_ ) snake_case_ = torch.nn.Linear(1 , 1 ) snake_case_ = accelerator.prepare(lowerCAmelCase_ ) snake_case_ = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) snake_case_ = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): snake_case_ = train_dl.batch_sampler.even_batches snake_case_ = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def _lowerCAmelCase ( )->Any: '''simple docstring''' snake_case_ = True snake_case_ = False snake_case_ = create_accelerator(even_batches=lowerCAmelCase_ ) snake_case_ = torch.nn.Linear(1 , 1 ) snake_case_ = accelerator.prepare(lowerCAmelCase_ ) create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase_ ) snake_case_ = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings("ignore" ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): snake_case_ = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def _lowerCAmelCase ( )->List[Any]: '''simple docstring''' snake_case_ = create_accelerator() snake_case_ = torch.nn.Linear(1 , 1 ) snake_case_ = accelerator.prepare(lowerCAmelCase_ ) create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase_ ) with warnings.catch_warnings(record=lowerCAmelCase_ ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): pass assert issubclass(w[-1].category , lowerCAmelCase_ ) assert "only supported for map-style datasets" in str(w[-1].message ) def _lowerCAmelCase ( )->str: '''simple docstring''' snake_case_ = create_accelerator() accelerator.print("Test that even_batches variable ensures uniform batches across processes" ) test_default_ensures_even_batch_sizes() accelerator.print("Run tests with even_batches disabled" ) test_can_disable_even_batches() accelerator.print("Test joining uneven inputs" ) test_can_join_uneven_inputs() accelerator.print("Test overriding even_batches when joining uneven inputs" ) test_join_can_override_even_batches() accelerator.print("Test overriding even_batches for mixed dataloader types" ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print("Test overriding even_batches raises a warning for iterable dataloaders" ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print("Test join with non DDP distributed raises warning" ) snake_case_ = accelerator.state.distributed_type snake_case_ = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(lowerCAmelCase_ ) snake_case_ = original_state if __name__ == "__main__": main()
159
from __future__ import annotations import os from typing import Any import requests SCREAMING_SNAKE_CASE :Tuple = '''https://api.github.com''' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user SCREAMING_SNAKE_CASE :Tuple = BASE_URL + '''/user''' # https://github.com/settings/tokens SCREAMING_SNAKE_CASE :Optional[Any] = os.environ.get('''USER_TOKEN''', '''''') def _lowerCAmelCase ( lowerCAmelCase_ :str )->dict[Any, Any]: '''simple docstring''' snake_case_ = { "Authorization": F'''token {auth_token}''', "Accept": "application/vnd.github.v3+json", } return requests.get(lowerCAmelCase_ , headers=lowerCAmelCase_ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F'''{key}: {value}''') else: raise ValueError('''\'USER_TOKEN\' field cannot be empty.''')
159
1
'''simple docstring''' from PIL import Image def snake_case_ ( __SCREAMING_SNAKE_CASE : Image , __SCREAMING_SNAKE_CASE : int ): """simple docstring""" lowercase_ : Optional[int] = (259 * (level + 255)) / (255 * (259 - level)) def contrast(__SCREAMING_SNAKE_CASE : int ) -> int: return int(128 + factor * (c - 128) ) return img.point(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change contrast to 170 _lowercase : Union[str, Any] = change_contrast(img, 1_7_0) cont_img.save("image_data/lena_high_contrast.png", format="png")
264
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class lowerCAmelCase__ : def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=10 , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=32 * 4 , __SCREAMING_SNAKE_CASE=32 * 6 , __SCREAMING_SNAKE_CASE=4 , __SCREAMING_SNAKE_CASE=32 , ): """simple docstring""" lowercase_ : Tuple = parent lowercase_ : Optional[int] = batch_size lowercase_ : Dict = is_training lowercase_ : Optional[Any] = use_auxiliary_loss lowercase_ : Optional[Any] = num_queries lowercase_ : Any = num_channels lowercase_ : str = min_size lowercase_ : str = max_size lowercase_ : Optional[Any] = num_labels lowercase_ : List[str] = mask_feature_size def _snake_case ( self ): """simple docstring""" lowercase_ : Dict = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( __SCREAMING_SNAKE_CASE ) lowercase_ : str = torch.ones([self.batch_size, self.min_size, self.max_size] , device=__SCREAMING_SNAKE_CASE ) lowercase_ : str = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=__SCREAMING_SNAKE_CASE ) > 0.5 ).float() lowercase_ : str = (torch.rand((self.batch_size, self.num_labels) , device=__SCREAMING_SNAKE_CASE ) > 0.5).long() lowercase_ : Optional[int] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _snake_case ( self ): """simple docstring""" return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_28 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def _snake_case ( self ): """simple docstring""" lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : Optional[Any] = self.prepare_config_and_inputs() lowercase_ : Any = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : List[str] = output.encoder_hidden_states lowercase_ : List[Any] = output.pixel_decoder_hidden_states lowercase_ : int = output.transformer_decoder_hidden_states self.parent.assertTrue(len(__SCREAMING_SNAKE_CASE ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__SCREAMING_SNAKE_CASE ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__SCREAMING_SNAKE_CASE ) , config.decoder_config.decoder_layers ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=False ): """simple docstring""" with torch.no_grad(): lowercase_ : Any = MaskFormerModel(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() lowercase_ : Dict = model(pixel_values=__SCREAMING_SNAKE_CASE , pixel_mask=__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = model(__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Dict = MaskFormerForInstanceSegmentation(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() def comm_check_on_output(__SCREAMING_SNAKE_CASE ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): lowercase_ : Tuple = model(pixel_values=__SCREAMING_SNAKE_CASE , pixel_mask=__SCREAMING_SNAKE_CASE ) lowercase_ : str = model(__SCREAMING_SNAKE_CASE ) comm_check_on_output(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = model( pixel_values=__SCREAMING_SNAKE_CASE , pixel_mask=__SCREAMING_SNAKE_CASE , mask_labels=__SCREAMING_SNAKE_CASE , class_labels=__SCREAMING_SNAKE_CASE ) comm_check_on_output(__SCREAMING_SNAKE_CASE ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowerCAmelCase__ ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCAmelCase_ = ( {'''feature-extraction''': MaskFormerModel, '''image-segmentation''': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def _snake_case ( self ): """simple docstring""" lowercase_ : int = MaskFormerModelTester(self ) lowercase_ : Tuple = ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , has_text_modality=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() def _snake_case ( self ): """simple docstring""" lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*__SCREAMING_SNAKE_CASE ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def _snake_case ( self ): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def _snake_case ( self ): """simple docstring""" pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _snake_case ( self ): """simple docstring""" pass def _snake_case ( self ): """simple docstring""" lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : Union[str, Any] = model_class(__SCREAMING_SNAKE_CASE ) lowercase_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : Dict = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE ) @slow def _snake_case ( self ): """simple docstring""" for model_name in ["facebook/maskformer-swin-small-coco"]: lowercase_ : Tuple = MaskFormerModel.from_pretrained(__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : Union[str, Any] = (self.model_tester.min_size,) * 2 lowercase_ : int = { '''pixel_values''': torch.randn((2, 3, *size) , device=__SCREAMING_SNAKE_CASE ), '''mask_labels''': torch.randn((2, 10, *size) , device=__SCREAMING_SNAKE_CASE ), '''class_labels''': torch.zeros(2 , 10 , device=__SCREAMING_SNAKE_CASE ).long(), } lowercase_ : Any = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = model(**__SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.loss is not None ) def _snake_case ( self ): """simple docstring""" lowercase_ , lowercase_ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ , lowercase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : List[str] = model_class(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = model(**__SCREAMING_SNAKE_CASE , output_attentions=__SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.attentions is not None ) def _snake_case ( self ): """simple docstring""" if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss lowercase_ : Optional[Any] = self.all_model_classes[1] lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() lowercase_ : Optional[int] = model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.train() lowercase_ : Any = model(__SCREAMING_SNAKE_CASE , mask_labels=__SCREAMING_SNAKE_CASE , class_labels=__SCREAMING_SNAKE_CASE ).loss loss.backward() def _snake_case ( self ): """simple docstring""" lowercase_ : Any = self.all_model_classes[1] lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : Any = self.model_tester.prepare_config_and_inputs() lowercase_ : Tuple = True lowercase_ : Optional[Any] = True lowercase_ : Tuple = model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.train() lowercase_ : List[Any] = model(__SCREAMING_SNAKE_CASE , mask_labels=__SCREAMING_SNAKE_CASE , class_labels=__SCREAMING_SNAKE_CASE ) lowercase_ : Any = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() lowercase_ : Any = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't lowercase_ : List[Any] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() lowercase_ : Tuple = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) _lowercase : int = 1E-4 def snake_case_ ( ): """simple docstring""" lowercase_ : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class lowerCAmelCase__ ( unittest.TestCase ): @cached_property def _snake_case ( self ): """simple docstring""" return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def _snake_case ( self ): """simple docstring""" lowercase_ : Any = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : Any = image_processor(__SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(__SCREAMING_SNAKE_CASE , (1, 3, 8_00, 10_88) ) with torch.no_grad(): lowercase_ : Any = model(**__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = torch.tensor( [[-0.0_482, 0.9_228, 0.4_951], [-0.2_547, 0.8_017, 0.8_527], [-0.0_069, 0.3_385, -0.0_089]] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) lowercase_ : List[Any] = torch.tensor( [[-0.8_422, -0.8_434, -0.9_718], [-1.0_144, -0.5_565, -0.4_195], [-1.0_038, -0.4_484, -0.1_961]] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) lowercase_ : Dict = torch.tensor( [[0.2_852, -0.0_159, 0.9_735], [0.6_254, 0.1_858, 0.8_529], [-0.0_680, -0.4_116, 1.8_413]] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) def _snake_case ( self ): """simple docstring""" lowercase_ : List[Any] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__SCREAMING_SNAKE_CASE ) .eval() ) lowercase_ : List[str] = self.default_image_processor lowercase_ : Union[str, Any] = prepare_img() lowercase_ : List[str] = image_processor(__SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(__SCREAMING_SNAKE_CASE , (1, 3, 8_00, 10_88) ) with torch.no_grad(): lowercase_ : Tuple = model(**__SCREAMING_SNAKE_CASE ) # masks_queries_logits lowercase_ : Union[str, Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowercase_ : Any = [ [-1.3_737_124, -1.7_724_937, -1.9_364_233], [-1.5_977_281, -1.9_867_939, -2.1_523_695], [-1.5_795_398, -1.9_269_832, -2.093_942], ] lowercase_ : List[str] = torch.tensor(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) # class_queries_logits lowercase_ : Optional[int] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowercase_ : Any = torch.tensor( [ [1.6_5_1_2E0_0, -5.2_5_7_2E0_0, -3.3_5_1_9E0_0], [3.6_1_6_9E-0_2, -5.9_0_2_5E0_0, -2.9_3_1_3E0_0], [1.0_7_6_6E-0_4, -7.7_6_3_0E0_0, -5.1_2_6_3E0_0], ] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) def _snake_case ( self ): """simple docstring""" lowercase_ : int = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(__SCREAMING_SNAKE_CASE ) .eval() ) lowercase_ : Tuple = self.default_image_processor lowercase_ : Any = prepare_img() lowercase_ : Union[str, Any] = image_processor(__SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(__SCREAMING_SNAKE_CASE , (1, 3, 8_00, 10_88) ) with torch.no_grad(): lowercase_ : Union[str, Any] = model(**__SCREAMING_SNAKE_CASE ) # masks_queries_logits lowercase_ : Union[str, Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowercase_ : Optional[int] = [[-0.9_046, -2.6_366, -4.6_062], [-3.4_179, -5.7_890, -8.8_057], [-4.9_179, -7.6_560, -10.7_711]] lowercase_ : Dict = torch.tensor(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) # class_queries_logits lowercase_ : Any = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowercase_ : Union[str, Any] = torch.tensor( [[4.7_188, -3.2_585, -2.8_857], [6.6_871, -2.9_181, -1.2_487], [7.2_449, -2.2_764, -2.1_874]] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=__SCREAMING_SNAKE_CASE ) ) def _snake_case ( self ): """simple docstring""" lowercase_ : int = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__SCREAMING_SNAKE_CASE ) .eval() ) lowercase_ : int = self.default_image_processor lowercase_ : Optional[Any] = image_processor( [np.zeros((3, 8_00, 13_33) ), np.zeros((3, 8_00, 13_33) )] , segmentation_maps=[np.zeros((3_84, 3_84) ).astype(np.floataa ), np.zeros((3_84, 3_84) ).astype(np.floataa )] , return_tensors='''pt''' , ) lowercase_ : Optional[int] = inputs['''pixel_values'''].to(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = [el.to(__SCREAMING_SNAKE_CASE ) for el in inputs['''mask_labels''']] lowercase_ : int = [el.to(__SCREAMING_SNAKE_CASE ) for el in inputs['''class_labels''']] with torch.no_grad(): lowercase_ : List[str] = model(**__SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.loss is not None )
264
1
'''simple docstring''' import requests def __lowercase ( __lowercase , __lowercase ) -> None: '''simple docstring''' _A = {"Content-Type": "application/json"} _A = requests.post(__lowercase , json={"text": message_body} , headers=__lowercase ) if response.status_code != 200: _A = ( "Request to slack returned an error " F'''{response.status_code}, the response is:\n{response.text}''' ) raise ValueError(__lowercase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('''<YOUR MESSAGE BODY>''', '''<SLACK CHANNEL URL>''')
79
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: List[str] ): '''simple docstring''' _snake_case : int = data _snake_case : Dict = [0X67452301, 0Xefcdab89, 0X98badcfe, 0X10325476, 0Xc3d2e1f0] @staticmethod def UpperCamelCase_ ( a_: Optional[Any], a_: Dict ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0Xffffffff def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _snake_case : Optional[int] = self.data + padding + struct.pack(""">Q""", 8 * len(self.data ) ) return padded_data def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0, len(self.padded_data ), 64 ) ] def UpperCamelCase_ ( self: Optional[Any], a_: List[Any] ): '''simple docstring''' _snake_case : List[str] = list(struct.unpack(""">16L""", a_ ) ) + [0] * 64 for i in range(16, 80 ): _snake_case : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]), 1 ) return w def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = self.padding() _snake_case : str = self.split_blocks() for block in self.blocks: _snake_case : Any = self.expand_block(a_ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = self.h for i in range(0, 80 ): if 0 <= i < 20: _snake_case : int = (b & c) | ((~b) & d) _snake_case : str = 0X5a827999 elif 20 <= i < 40: _snake_case : Optional[int] = b ^ c ^ d _snake_case : str = 0X6ed9eba1 elif 40 <= i < 60: _snake_case : List[Any] = (b & c) | (b & d) | (c & d) _snake_case : List[Any] = 0X8f1bbcdc elif 60 <= i < 80: _snake_case : List[Any] = b ^ c ^ d _snake_case : int = 0Xca62c1d6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = ( self.rotate(a_, 5 ) + f + e + k + expanded_block[i] & 0Xffffffff, a, self.rotate(a_, 30 ), c, d, ) _snake_case : Union[str, Any] = ( self.h[0] + a & 0Xffffffff, self.h[1] + b & 0Xffffffff, self.h[2] + c & 0Xffffffff, self.h[3] + d & 0Xffffffff, self.h[4] + e & 0Xffffffff, ) return ("{:08x}" * 5).format(*self.h ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = B"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) _snake_case : Union[str, Any] = parser.parse_args() _snake_case : List[Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: _snake_case : str = f.read() else: _snake_case : int = bytes(snake_case__ , """utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
64
0
import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline _UpperCAmelCase : Optional[int] = version.parse(version.parse(torch.__version__).base_version) < version.parse("1.11") def A ( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=False , ) -> Optional[Any]: '''simple docstring''' output_path.parent.mkdir(parents=lowercase , exist_ok=lowercase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowercase , lowercase , f=output_path.as_posix() , input_names=lowercase , output_names=lowercase , dynamic_axes=lowercase , do_constant_folding=lowercase , use_external_data_format=lowercase , enable_onnx_checker=lowercase , opset_version=lowercase , ) else: export( lowercase , lowercase , f=output_path.as_posix() , input_names=lowercase , output_names=lowercase , dynamic_axes=lowercase , do_constant_folding=lowercase , opset_version=lowercase , ) @torch.no_grad() def A ( lowercase , lowercase , lowercase , lowercase = False ) -> Any: '''simple docstring''' UpperCamelCase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): UpperCamelCase = 'cuda' elif fpaa and not torch.cuda.is_available(): raise ValueError('`float16` model export is only supported on GPUs with CUDA' ) else: UpperCamelCase = 'cpu' UpperCamelCase = StableDiffusionPipeline.from_pretrained(lowercase , torch_dtype=lowercase ).to(lowercase ) UpperCamelCase = Path(lowercase ) # TEXT ENCODER UpperCamelCase = pipeline.text_encoder.config.max_position_embeddings UpperCamelCase = pipeline.text_encoder.config.hidden_size UpperCamelCase = pipeline.tokenizer( 'A sample prompt' , padding='max_length' , max_length=pipeline.tokenizer.model_max_length , truncation=lowercase , return_tensors='pt' , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=lowercase , dtype=torch.intaa )) , output_path=output_path / 'text_encoder' / 'model.onnx' , ordered_input_names=['input_ids'] , output_names=['last_hidden_state', 'pooler_output'] , dynamic_axes={ 'input_ids': {0: 'batch', 1: 'sequence'}, } , opset=lowercase , ) del pipeline.text_encoder # UNET UpperCamelCase = pipeline.unet.config.in_channels UpperCamelCase = pipeline.unet.config.sample_size UpperCamelCase = output_path / 'unet' / 'model.onnx' onnx_export( pipeline.unet , model_args=( torch.randn(2 , lowercase , lowercase , lowercase ).to(device=lowercase , dtype=lowercase ), torch.randn(2 ).to(device=lowercase , dtype=lowercase ), torch.randn(2 , lowercase , lowercase ).to(device=lowercase , dtype=lowercase ), False, ) , output_path=lowercase , ordered_input_names=['sample', 'timestep', 'encoder_hidden_states', 'return_dict'] , output_names=['out_sample'] , dynamic_axes={ 'sample': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, 'timestep': {0: 'batch'}, 'encoder_hidden_states': {0: 'batch', 1: 'sequence'}, } , opset=lowercase , use_external_data_format=lowercase , ) UpperCamelCase = str(unet_path.absolute().as_posix() ) UpperCamelCase = os.path.dirname(lowercase ) UpperCamelCase = onnx.load(lowercase ) # clean up existing tensor files shutil.rmtree(lowercase ) os.mkdir(lowercase ) # collate external tensor files into one onnx.save_model( lowercase , lowercase , save_as_external_data=lowercase , all_tensors_to_one_file=lowercase , location='weights.pb' , convert_attribute=lowercase , ) del pipeline.unet # VAE ENCODER UpperCamelCase = pipeline.vae UpperCamelCase = vae_encoder.config.in_channels UpperCamelCase = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder UpperCamelCase = lambda lowercase , lowercase : vae_encoder.encode(lowercase , lowercase )[0].sample() onnx_export( lowercase , model_args=( torch.randn(1 , lowercase , lowercase , lowercase ).to(device=lowercase , dtype=lowercase ), False, ) , output_path=output_path / 'vae_encoder' / 'model.onnx' , ordered_input_names=['sample', 'return_dict'] , output_names=['latent_sample'] , dynamic_axes={ 'sample': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, } , opset=lowercase , ) # VAE DECODER UpperCamelCase = pipeline.vae UpperCamelCase = vae_decoder.config.latent_channels UpperCamelCase = vae_decoder.config.out_channels # forward only through the decoder part UpperCamelCase = vae_encoder.decode onnx_export( lowercase , model_args=( torch.randn(1 , lowercase , lowercase , lowercase ).to(device=lowercase , dtype=lowercase ), False, ) , output_path=output_path / 'vae_decoder' / 'model.onnx' , ordered_input_names=['latent_sample', 'return_dict'] , output_names=['sample'] , dynamic_axes={ 'latent_sample': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, } , opset=lowercase , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: UpperCamelCase = pipeline.safety_checker UpperCamelCase = safety_checker.config.vision_config.num_channels UpperCamelCase = safety_checker.config.vision_config.image_size UpperCamelCase = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , lowercase , lowercase , lowercase , ).to(device=lowercase , dtype=lowercase ), torch.randn(1 , lowercase , lowercase , lowercase ).to(device=lowercase , dtype=lowercase ), ) , output_path=output_path / 'safety_checker' / 'model.onnx' , ordered_input_names=['clip_input', 'images'] , output_names=['out_images', 'has_nsfw_concepts'] , dynamic_axes={ 'clip_input': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, 'images': {0: 'batch', 1: 'height', 2: 'width', 3: 'channels'}, } , opset=lowercase , ) del pipeline.safety_checker UpperCamelCase = OnnxRuntimeModel.from_pretrained(output_path / 'safety_checker' ) UpperCamelCase = pipeline.feature_extractor else: UpperCamelCase = None UpperCamelCase = None UpperCamelCase = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / 'vae_encoder' ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / 'vae_decoder' ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / 'text_encoder' ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / 'unet' ) , scheduler=pipeline.scheduler , safety_checker=lowercase , feature_extractor=lowercase , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(lowercase ) print('ONNX pipeline saved to' , lowercase ) del pipeline del onnx_pipeline UpperCamelCase = OnnxStableDiffusionPipeline.from_pretrained(lowercase , provider='CPUExecutionProvider' ) print('ONNX pipeline is loadable' ) if __name__ == "__main__": _UpperCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( "--model_path", type=str, required=True, help="Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).", ) parser.add_argument("--output_path", type=str, required=True, help="Path to the output model.") parser.add_argument( "--opset", default=14, type=int, help="The version of the ONNX operator set to use.", ) parser.add_argument("--fp16", action="store_true", default=False, help="Export the models in `float16` mode") _UpperCAmelCase : List[str] = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
368
from __future__ import annotations def A ( lowercase , lowercase ) -> tuple[int, int]: '''simple docstring''' if b == 0: return (1, 0) ((UpperCamelCase) , (UpperCamelCase)) = extended_euclid(lowercase , a % b ) UpperCamelCase = a // b return (y, x - k * y) def A ( lowercase , lowercase , lowercase , lowercase ) -> int: '''simple docstring''' ((UpperCamelCase) , (UpperCamelCase)) = extended_euclid(lowercase , lowercase ) UpperCamelCase = na * na UpperCamelCase = ra * x * na + ra * y * na return (n % m + m) % m def A ( lowercase , lowercase ) -> int: '''simple docstring''' ((UpperCamelCase) , (UpperCamelCase)) = extended_euclid(lowercase , lowercase ) if b < 0: UpperCamelCase = (b % n + n) % n return b def A ( lowercase , lowercase , lowercase , lowercase ) -> int: '''simple docstring''' UpperCamelCase , UpperCamelCase = invert_modulo(lowercase , lowercase ), invert_modulo(lowercase , lowercase ) UpperCamelCase = na * na UpperCamelCase = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name="chinese_remainder_theorem", verbose=True) testmod(name="chinese_remainder_theorem2", verbose=True) testmod(name="invert_modulo", verbose=True) testmod(name="extended_euclid", verbose=True)
110
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowercase__ : Union[str, Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE (a__ ): def __init__( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' warnings.warn( 'The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use LayoutLMv2ImageProcessor instead.' , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase)
190
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ : int = logging.get_logger(__name__) lowercase__ : int = { '''andreasmadsen/efficient_mlm_m0.40''': ( '''https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''roberta-prelayernorm''' def __init__( self , _UpperCAmelCase=5_0265 , _UpperCAmelCase=768 , _UpperCAmelCase=12 , _UpperCAmelCase=12 , _UpperCAmelCase=3072 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=1e-1_2 , _UpperCAmelCase=1 , _UpperCAmelCase=0 , _UpperCAmelCase=2 , _UpperCAmelCase="absolute" , _UpperCAmelCase=True , _UpperCAmelCase=None , **_UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase) __A : Optional[int] = vocab_size __A : List[Any] = hidden_size __A : Optional[int] = num_hidden_layers __A : Optional[Any] = num_attention_heads __A : List[str] = hidden_act __A : Dict = intermediate_size __A : Optional[int] = hidden_dropout_prob __A : Optional[Any] = attention_probs_dropout_prob __A : Tuple = max_position_embeddings __A : Union[str, Any] = type_vocab_size __A : Any = initializer_range __A : str = layer_norm_eps __A : int = position_embedding_type __A : Optional[Any] = use_cache __A : Any = classifier_dropout class SCREAMING_SNAKE_CASE (a__ ): @property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' if self.task == "multiple-choice": __A : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __A : Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
190
1
# Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def A_ ( A__ , A__ , A__ , A__ ) -> List[str]: a__ : Dict = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, nicht wahr?', } # BLUE scores as follows: # "pair": [fairseq, transformers] a__ : Optional[Any] = { 'wmt16-en-de-dist-12-1': [28.3, 27.52], 'wmt16-en-de-dist-6-1': [27.4, 27.11], 'wmt16-en-de-12-1': [26.9, 25.75], } a__ : Any = F'{src_lang}-{tgt_lang}' a__ : Optional[int] = F'\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "allenai/{model_name}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don\'t use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n' model_card_dir.mkdir(parents=A__ , exist_ok=A__ ) a__ : Union[str, Any] = os.path.join(A__ , 'README.md' ) print(F'Generating {path}' ) with open(A__ , 'w' , encoding='utf-8' ) as f: f.write(A__ ) # make sure we are under the root of the project lowercase : Dict = Path(__file__).resolve().parent.parent.parent lowercase : Tuple = repo_dir / """model_cards""" for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: lowercase : Any = model_cards_dir / """allenai""" / model_name write_model_card(model_card_dir, src_lang="""en""", tgt_lang="""de""", model_name=model_name)
225
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def A_ ( A__ ) -> str: a__ : Any = 384 if "tiny" in model_name: a__ : List[Any] = [3, 3, 9, 3] a__ : Optional[Any] = [96, 192, 384, 768] if "small" in model_name: a__ : Union[str, Any] = [3, 3, 27, 3] a__ : List[Any] = [96, 192, 384, 768] if "base" in model_name: a__ : int = [3, 3, 27, 3] a__ : List[str] = [128, 256, 512, 1024] a__ : Optional[int] = 512 if "large" in model_name: a__ : Optional[int] = [3, 3, 27, 3] a__ : Any = [192, 384, 768, 1536] a__ : int = 768 if "xlarge" in model_name: a__ : str = [3, 3, 27, 3] a__ : int = [256, 512, 1024, 2048] a__ : List[str] = 1024 # set label information a__ : int = 150 a__ : List[Any] = 'huggingface/label-files' a__ : str = 'ade20k-id2label.json' a__ : Optional[int] = json.load(open(hf_hub_download(A__ , A__ , repo_type='dataset' ) , 'r' ) ) a__ : List[str] = {int(A__ ): v for k, v in idalabel.items()} a__ : Union[str, Any] = {v: k for k, v in idalabel.items()} a__ : List[Any] = ConvNextConfig( depths=A__ , hidden_sizes=A__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) a__ : Optional[int] = UperNetConfig( backbone_config=A__ , auxiliary_in_channels=A__ , num_labels=A__ , idalabel=A__ , labelaid=A__ , ) return config def A_ ( A__ ) -> Tuple: a__ : Optional[int] = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.{j}.gamma', F'backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.weight', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.bias', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.weight', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.bias', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias') ) if i > 0: rename_keys.append((F'backbone.downsample_layers.{i}.0.weight', F'backbone.encoder.stages.{i}.downsampling_layer.0.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.0.bias', F'backbone.encoder.stages.{i}.downsampling_layer.0.bias') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.weight', F'backbone.encoder.stages.{i}.downsampling_layer.1.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.bias', F'backbone.encoder.stages.{i}.downsampling_layer.1.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def A_ ( A__ , A__ , A__ ) -> str: a__ : List[str] = dct.pop(A__ ) a__ : int = val def A_ ( A__ , A__ , A__ ) -> str: a__ : Tuple = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } a__ : Dict = model_name_to_url[model_name] a__ : Optional[int] = torch.hub.load_state_dict_from_url(A__ , map_location='cpu' )['state_dict'] a__ : List[Any] = get_upernet_config(A__ ) a__ : Dict = UperNetForSemanticSegmentation(A__ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): a__ : Dict = state_dict.pop(A__ ) if "bn" in key: a__ : Optional[int] = key.replace('bn' , 'batch_norm' ) a__ : List[Any] = val # rename keys a__ : Union[str, Any] = create_rename_keys(A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) model.load_state_dict(A__ ) # verify on image a__ : str = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' a__ : int = Image.open(requests.get(A__ , stream=A__ ).raw ).convert('RGB' ) a__ : Union[str, Any] = SegformerImageProcessor() a__ : Union[str, Any] = processor(A__ , return_tensors='pt' ).pixel_values with torch.no_grad(): a__ : Optional[Any] = model(A__ ) if model_name == "upernet-convnext-tiny": a__ : Union[str, Any] = torch.tensor( [[-8.81_10, -8.81_10, -8.65_21], [-8.81_10, -8.81_10, -8.65_21], [-8.77_46, -8.77_46, -8.61_30]] ) elif model_name == "upernet-convnext-small": a__ : int = torch.tensor( [[-8.82_36, -8.82_36, -8.67_71], [-8.82_36, -8.82_36, -8.67_71], [-8.76_38, -8.76_38, -8.62_40]] ) elif model_name == "upernet-convnext-base": a__ : int = torch.tensor( [[-8.85_58, -8.85_58, -8.69_05], [-8.85_58, -8.85_58, -8.69_05], [-8.76_69, -8.76_69, -8.60_21]] ) elif model_name == "upernet-convnext-large": a__ : Optional[Any] = torch.tensor( [[-8.66_60, -8.66_60, -8.62_10], [-8.66_60, -8.66_60, -8.62_10], [-8.63_10, -8.63_10, -8.59_64]] ) elif model_name == "upernet-convnext-xlarge": a__ : Optional[int] = torch.tensor( [[-8.49_80, -8.49_80, -8.39_77], [-8.49_80, -8.49_80, -8.39_77], [-8.43_79, -8.43_79, -8.34_12]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , A__ , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(A__ ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": lowercase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""upernet-convnext-tiny""", type=str, choices=[F"""upernet-convnext-{size}""" for size in ["""tiny""", """small""", """base""", """large""", """xlarge"""]], help="""Name of the ConvNext UperNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowercase : str = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
225
1
'''simple docstring''' import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def __UpperCamelCase ( lowercase__ : str, lowercase__ : str, **lowercase__ : Dict ): '''simple docstring''' __lowercase =AutoConfig.from_pretrained(_snake_case, **_snake_case ) __lowercase =AutoModelForSeqaSeqLM.from_config(_snake_case ) model.save_pretrained(_snake_case ) AutoTokenizer.from_pretrained(_snake_case ).save_pretrained(_snake_case ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
141
"""simple docstring""" from __future__ import annotations def _snake_case ( _snake_case : int , _snake_case : int ) -> list[list[int]]: '''simple docstring''' _A = [] create_all_state(1 , _snake_case , _snake_case , [] , _snake_case ) return result def _snake_case ( _snake_case : int , _snake_case : int , _snake_case : int , _snake_case : list[int] , _snake_case : list[list[int]] , ) -> None: '''simple docstring''' if level == 0: total_list.append(current_list[:] ) return for i in range(_snake_case , total_number - level + 2 ): current_list.append(_snake_case ) create_all_state(i + 1 , _snake_case , level - 1 , _snake_case , _snake_case ) current_list.pop() def _snake_case ( _snake_case : list[list[int]] ) -> None: '''simple docstring''' for i in total_list: print(*_snake_case ) if __name__ == "__main__": a = 4 a = 2 a = generate_all_combinations(n, k) print_all_state(total_list)
315
0
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class UpperCamelCase__ : _SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] _SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] _SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] _SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] _SCREAMING_SNAKE_CASE : int _SCREAMING_SNAKE_CASE : int _SCREAMING_SNAKE_CASE : float _SCREAMING_SNAKE_CASE : float _SCREAMING_SNAKE_CASE : Tuple[int] def lowerCAmelCase (self : Optional[Any] ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def lowerCAmelCase (self : List[Any] ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def lowerCAmelCase (self : List[Any] ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def lowerCAmelCase (self : Any ): __a : str = torch.arange(self.height * self.width ) __a : int = torch.stack( [ pixel_indices % self.width, torch.div(snake_case_ , self.width , rounding_mode='''trunc''' ), ] , axis=1 , ) return coords @property def lowerCAmelCase (self : List[str] ): __a , *__a : Any = self.shape __a : Tuple = int(np.prod(snake_case_ ) ) __a : int = self.get_image_coords() __a : Optional[int] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a : Union[str, Any] = self.get_camera_rays(snake_case_ ) __a : Tuple = rays.view(snake_case_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def lowerCAmelCase (self : Union[str, Any] , snake_case_ : torch.Tensor ): __a , *__a , __a : Optional[int] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a : Optional[int] = coords.view(snake_case_ , -1 , 2 ) __a : Dict = self.resolution() __a : List[Any] = self.fov() __a : Any = (flat.float() / (res - 1)) * 2 - 1 __a : Any = fracs * torch.tan(fov / 2 ) __a : Tuple = fracs.view(snake_case_ , -1 , 2 ) __a : int = ( self.z.view(snake_case_ , 1 , 3 ) + self.x.view(snake_case_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(snake_case_ , 1 , 3 ) * fracs[:, :, 1:] ) __a : Optional[Any] = directions / directions.norm(dim=-1 , keepdim=snake_case_ ) __a : List[Any] = torch.stack( [ torch.broadcast_to(self.origin.view(snake_case_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(snake_case_ , *snake_case_ , 2 , 3 ) def lowerCAmelCase (self : List[str] , snake_case_ : int , snake_case_ : int ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=snake_case_ , height=snake_case_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __UpperCamelCase ( lowerCAmelCase__ : int ): __a : List[str] = [] __a : Optional[Any] = [] __a : Any = [] __a : List[str] = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a : Optional[Any] = np.array([np.sin(lowerCAmelCase__ ), np.cos(lowerCAmelCase__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a : str = -z * 4 __a : Tuple = np.array([np.cos(lowerCAmelCase__ ), -np.sin(lowerCAmelCase__ ), 0.0] ) __a : Tuple = np.cross(lowerCAmelCase__ , lowerCAmelCase__ ) origins.append(lowerCAmelCase__ ) xs.append(lowerCAmelCase__ ) ys.append(lowerCAmelCase__ ) zs.append(lowerCAmelCase__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(lowerCAmelCase__ , axis=0 ) ).float() , x=torch.from_numpy(np.stack(lowerCAmelCase__ , axis=0 ) ).float() , y=torch.from_numpy(np.stack(lowerCAmelCase__ , axis=0 ) ).float() , z=torch.from_numpy(np.stack(lowerCAmelCase__ , axis=0 ) ).float() , width=lowerCAmelCase__ , height=lowerCAmelCase__ , x_fov=0.7 , y_fov=0.7 , shape=(1, len(lowerCAmelCase__ )) , )
90
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ ={ 'configuration_efficientnet': [ 'EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EfficientNetConfig', 'EfficientNetOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =['EfficientNetImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'EfficientNetForImageClassification', 'EfficientNetModel', 'EfficientNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys lowercase__ =_LazyModule(__name__, globals()['__file__'], _import_structure)
90
1
"""simple docstring""" from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class lowerCamelCase : lowerCamelCase__ : List[str] lowerCamelCase__ : Optional[str] = None # Automatically constructed lowerCamelCase__ : ClassVar[str] = "dict" lowerCamelCase__ : ClassVar[Any] = None lowerCamelCase__ : str = field(default='Translation' ,init=_UpperCamelCase ,repr=_UpperCamelCase ) def __call__( self : Any ) -> Union[str, Any]: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value("""string""" ) for k in sorted(self.languages )} @dataclass class lowerCamelCase : lowerCamelCase__ : Optional[List] = None lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : Optional[str] = None # Automatically constructed lowerCamelCase__ : ClassVar[str] = "dict" lowerCamelCase__ : ClassVar[Any] = None lowerCamelCase__ : str = field(default='TranslationVariableLanguages' ,init=_UpperCamelCase ,repr=_UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: SCREAMING_SNAKE_CASE__ = sorted(set(self.languages ) ) if self.languages else None SCREAMING_SNAKE_CASE__ = len(self.languages ) if self.languages else None def __call__( self : str ) -> Dict: return pa.struct({"""language""": pa.list_(pa.string() ), """translation""": pa.list_(pa.string() )} ) def SCREAMING_SNAKE_CASE ( self : int , __UpperCAmelCase : Tuple ) -> Dict: SCREAMING_SNAKE_CASE__ = set(self.languages ) if self.languages and set(__lowerCamelCase ) - lang_set: raise ValueError( F"""Some languages in example ({", ".join(sorted(set(__lowerCamelCase ) - lang_set ) )}) are not in valid set ({", ".join(__lowerCamelCase )}).""" ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. SCREAMING_SNAKE_CASE__ = [] for lang, text in translation_dict.items(): if isinstance(__lowerCamelCase , __lowerCamelCase ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = zip(*sorted(__lowerCamelCase ) ) return {"language": languages, "translation": translations} def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value("""string""" ) ), "translation": Sequence(Value("""string""" ) ), }
165
from __future__ import annotations from scipy.special import comb # type: ignore class snake_case__ : """simple docstring""" def __init__( self : Any , __lowerCamelCase : list[tuple[float, float]] ) -> Tuple: a = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. a = len(__lowerCamelCase ) - 1 def __UpperCAmelCase ( self : Union[str, Any] , __lowerCamelCase : float ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __lowerCamelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__lowerCamelCase ) , 5 ) == 1 return output_values def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = self.basis_function(__lowerCamelCase ) a = 0.0 a = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float = 0.01 ) -> List[str]: from matplotlib import pyplot as plt # type: ignore a = [] # x coordinates of points to plot a = [] # y coordinates of points to plot a = 0.0 while t <= 1: a = self.bezier_curve_function(__lowerCamelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size a = [i[0] for i in self.list_of_points] a = [i[1] for i in self.list_of_points] plt.plot( __lowerCamelCase , __lowerCamelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__lowerCamelCase , __lowerCamelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
107
0
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """encoder.layer_norm_for_extract""": """layer_norm_for_extract""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """label_embs_concat""": """label_embeddings_concat""", """mask_emb""": """masked_spec_embed""", """spk_proj""": """speaker_proj""", } _lowerCamelCase = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", """label_embeddings_concat""", """speaker_proj""", """layer_norm_for_extract""", ] def a__ ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Any ) -> List[str]: """simple docstring""" for attribute in key.split("." ): UpperCAmelCase_ : Optional[Any] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: UpperCAmelCase_ : List[Any] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: UpperCAmelCase_ : Any = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": UpperCAmelCase_ : List[str] = value elif weight_type == "weight_g": UpperCAmelCase_ : Any = value elif weight_type == "weight_v": UpperCAmelCase_ : int = value elif weight_type == "bias": UpperCAmelCase_ : List[Any] = value else: UpperCAmelCase_ : Tuple = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def a__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Any ) -> int: """simple docstring""" UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : List[str] = fairseq_model.state_dict() UpperCAmelCase_ : Any = hf_model.unispeech_sat.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase_ : int = False if "conv_layers" in name: load_conv_layer( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == "group" , ) UpperCAmelCase_ : Tuple = True else: for key, mapped_key in MAPPING.items(): UpperCAmelCase_ : List[Any] = "unispeech_sat." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: if "layer_norm_for_extract" in name and (".".join(name.split("." )[:-1] ) != key): # special case since naming is very similar continue UpperCAmelCase_ : Union[str, Any] = True if "*" in mapped_key: UpperCAmelCase_ : Optional[Any] = name.split(_SCREAMING_SNAKE_CASE )[0].split("." )[-2] UpperCAmelCase_ : str = mapped_key.replace("*" , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: UpperCAmelCase_ : Optional[Any] = "weight_g" elif "weight_v" in name: UpperCAmelCase_ : Optional[Any] = "weight_v" elif "bias" in name: UpperCAmelCase_ : Union[str, Any] = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase_ : List[Any] = "weight" else: UpperCAmelCase_ : int = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'''Unused weights: {unused_weights}''' ) def a__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase_ : Tuple = full_name.split("conv_layers." )[-1] UpperCAmelCase_ : Optional[Any] = name.split("." ) UpperCAmelCase_ : List[Any] = int(items[0] ) UpperCAmelCase_ : List[str] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) UpperCAmelCase_ : int = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) UpperCAmelCase_ : Union[str, Any] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.bias.data.shape} was found.''' ) UpperCAmelCase_ : Optional[int] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) UpperCAmelCase_ : List[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) @torch.no_grad() def a__ ( _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : List[str]=None , _SCREAMING_SNAKE_CASE : int=None , _SCREAMING_SNAKE_CASE : str=True ) -> Tuple: """simple docstring""" if config_path is not None: UpperCAmelCase_ : int = UniSpeechSatConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ : List[str] = UniSpeechSatConfig() UpperCAmelCase_ : List[str] = "" if is_finetuned: UpperCAmelCase_ : Optional[int] = UniSpeechSatForCTC(_SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ : List[Any] = UniSpeechSatForPreTraining(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) UpperCAmelCase_ : Optional[Any] = model[0].eval() recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) _lowerCamelCase = parser.parse_args() convert_unispeech_sat_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
360
'''simple docstring''' from __future__ import annotations from collections.abc import Sequence from typing import Literal def a__ ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str ) -> str | Literal[False]: """simple docstring""" UpperCAmelCase_ : Tuple = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Dict = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase_ : List[str] = "_" if count > 1: return False else: return "".join(_SCREAMING_SNAKE_CASE ) def a__ ( _SCREAMING_SNAKE_CASE : list[str] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : List[str] = [] while True: UpperCAmelCase_ : Any = ["$"] * len(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = [] for i in range(len(_SCREAMING_SNAKE_CASE ) ): for j in range(i + 1 , len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : Optional[int] = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase_ : Union[str, Any] = "*" UpperCAmelCase_ : List[Any] = "*" temp.append("X" ) for i in range(len(_SCREAMING_SNAKE_CASE ) ): if checka[i] == "$": pi.append(binary[i] ) if len(_SCREAMING_SNAKE_CASE ) == 0: return pi UpperCAmelCase_ : str = list(set(_SCREAMING_SNAKE_CASE ) ) def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Sequence[float] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : Optional[Any] = [] for minterm in minterms: UpperCAmelCase_ : Optional[Any] = "" for _ in range(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : Dict = str(minterm % 2 ) + string minterm //= 2 temp.append(_SCREAMING_SNAKE_CASE ) return temp def a__ ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int ) -> bool: """simple docstring""" UpperCAmelCase_ : Tuple = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[Any] = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def a__ ( _SCREAMING_SNAKE_CASE : list[list[int]] , _SCREAMING_SNAKE_CASE : list[str] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : Optional[int] = [] UpperCAmelCase_ : str = [0] * len(_SCREAMING_SNAKE_CASE ) for i in range(len(chart[0] ) ): UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Optional[int] = -1 for j in range(len(_SCREAMING_SNAKE_CASE ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase_ : Any = j if count == 1: UpperCAmelCase_ : Any = 1 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : int = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : List[str] = -1 UpperCAmelCase_ : int = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : str = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase_ : List[str] = count_n UpperCAmelCase_ : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : int = 0 def a__ ( _SCREAMING_SNAKE_CASE : list[str] , _SCREAMING_SNAKE_CASE : list[str] ) -> list[list[int]]: """simple docstring""" UpperCAmelCase_ : Dict = [[0 for x in range(len(_SCREAMING_SNAKE_CASE ) )] for x in range(len(_SCREAMING_SNAKE_CASE ) )] for i in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : Optional[int] = prime_implicants[i].count("_" ) for j in range(len(_SCREAMING_SNAKE_CASE ) ): if is_for_table(prime_implicants[i] , binary[j] , _SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : str = 1 return chart def a__ ( ) -> None: """simple docstring""" UpperCAmelCase_ : List[Any] = int(input("Enter the no. of variables\n" ) ) UpperCAmelCase_ : Tuple = [ float(_SCREAMING_SNAKE_CASE ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] UpperCAmelCase_ : int = decimal_to_binary(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = check(_SCREAMING_SNAKE_CASE ) print("Prime Implicants are:" ) print(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = prime_implicant_chart(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = selection(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print("Essential Prime Implicants are:" ) print(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() main()
67
0
import numpy as np def a__ ( A_ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def a__ ( A_ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
88
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ ): '''simple docstring''' _UpperCAmelCase = torch.load(snake_case_ , map_location="cpu" ) _UpperCAmelCase = chkpt["model"] # We have the base model one level deeper than the original XLM repository _UpperCAmelCase = {} for k, v in state_dict.items(): if "pred_layer" in k: _UpperCAmelCase = v else: _UpperCAmelCase = v _UpperCAmelCase = chkpt["params"] _UpperCAmelCase = {n: v for n, v in config.items() if not isinstance(snake_case_ , (torch.FloatTensor, numpy.ndarray) )} _UpperCAmelCase = chkpt["dico_word2id"] _UpperCAmelCase = {s + "</w>" if s.find("@@" ) == -1 and i > 13 else s.replace("@@" , "" ): i for s, i in vocab.items()} # Save pytorch-model _UpperCAmelCase = pytorch_dump_folder_path + "/" + WEIGHTS_NAME _UpperCAmelCase = pytorch_dump_folder_path + "/" + CONFIG_NAME _UpperCAmelCase = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(f"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(snake_case_ , snake_case_ ) print(f"""Save configuration file to {pytorch_config_dump_path}""" ) with open(snake_case_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(snake_case_ , indent=2 ) + "\n" ) print(f"""Save vocab file to {pytorch_config_dump_path}""" ) with open(snake_case_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(snake_case_ , indent=2 ) + "\n" ) if __name__ == "__main__": lowercase_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowercase_ : Optional[int] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
133
0
def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise ValueError("""multiplicative_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""multiplicative_persistence() does not accept negative values""" ) UpperCamelCase :Tuple = 0 UpperCamelCase :Optional[int] = str(__magic_name__ ) while len(__magic_name__ ) != 1: UpperCamelCase :Union[str, Any] = [int(__magic_name__ ) for i in num_string] UpperCamelCase :Tuple = 1 for i in range(0 , len(__magic_name__ ) ): total *= numbers[i] UpperCamelCase :List[str] = str(__magic_name__ ) steps += 1 return steps def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise ValueError("""additive_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""additive_persistence() does not accept negative values""" ) UpperCamelCase :List[Any] = 0 UpperCamelCase :int = str(__magic_name__ ) while len(__magic_name__ ) != 1: UpperCamelCase :Any = [int(__magic_name__ ) for i in num_string] UpperCamelCase :Optional[int] = 0 for i in range(0 , len(__magic_name__ ) ): total += numbers[i] UpperCamelCase :List[str] = str(__magic_name__ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
360
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Union[str, Any] = """char""" snake_case__ : Optional[int] = """bpe""" snake_case__ : Dict = """wp""" UpperCAmelCase_ : List[Any] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : List[Any] = ["""image_processor""", """char_tokenizer"""] snake_case__ : Dict = """ViTImageProcessor""" snake_case__ : List[str] = """MgpstrTokenizer""" def __init__( self : Optional[int] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Dict=None , **__lowerCamelCase : Any ): UpperCamelCase :Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __lowerCamelCase , ) UpperCamelCase :Optional[int] = kwargs.pop("""feature_extractor""" ) UpperCamelCase :List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) UpperCamelCase :Optional[int] = tokenizer UpperCamelCase :int = AutoTokenizer.from_pretrained("""gpt2""" ) UpperCamelCase :int = AutoTokenizer.from_pretrained("""bert-base-uncased""" ) super().__init__(__lowerCamelCase , __lowerCamelCase ) def __call__( self : str , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Dict=None , __lowerCamelCase : str=None , **__lowerCamelCase : Dict ): if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: UpperCamelCase :Tuple = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is not None: UpperCamelCase :Any = self.char_tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is None: return inputs elif images is None: return encodings else: UpperCamelCase :Dict = encodings["""input_ids"""] return inputs def _A ( self : Tuple , __lowerCamelCase : str ): UpperCamelCase , UpperCamelCase , UpperCamelCase :int = sequences UpperCamelCase :Tuple = char_preds.size(0 ) UpperCamelCase , UpperCamelCase :str = self._decode_helper(__lowerCamelCase , """char""" ) UpperCamelCase , UpperCamelCase :List[Any] = self._decode_helper(__lowerCamelCase , """bpe""" ) UpperCamelCase , UpperCamelCase :List[Any] = self._decode_helper(__lowerCamelCase , """wp""" ) UpperCamelCase :Any = [] UpperCamelCase :str = [] for i in range(__lowerCamelCase ): UpperCamelCase :Union[str, Any] = [char_scores[i], bpe_scores[i], wp_scores[i]] UpperCamelCase :Any = [char_strs[i], bpe_strs[i], wp_strs[i]] UpperCamelCase :str = scores.index(max(__lowerCamelCase ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) UpperCamelCase :Optional[Any] = {} UpperCamelCase :Dict = final_strs UpperCamelCase :Union[str, Any] = final_scores UpperCamelCase :List[str] = char_strs UpperCamelCase :Tuple = bpe_strs UpperCamelCase :Optional[Any] = wp_strs return out def _A ( self : int , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str] ): if format == DecodeType.CHARACTER: UpperCamelCase :List[str] = self.char_decode UpperCamelCase :Union[str, Any] = 1 UpperCamelCase :Optional[Any] = """[s]""" elif format == DecodeType.BPE: UpperCamelCase :Union[str, Any] = self.bpe_decode UpperCamelCase :str = 2 UpperCamelCase :int = """#""" elif format == DecodeType.WORDPIECE: UpperCamelCase :int = self.wp_decode UpperCamelCase :Any = 102 UpperCamelCase :int = """[SEP]""" else: raise ValueError(F"""Format {format} is not supported.""" ) UpperCamelCase , UpperCamelCase :int = [], [] UpperCamelCase :Any = pred_logits.size(0 ) UpperCamelCase :List[Any] = pred_logits.size(1 ) UpperCamelCase , UpperCamelCase :Optional[int] = pred_logits.topk(1 , dim=-1 , largest=__lowerCamelCase , sorted=__lowerCamelCase ) UpperCamelCase :Optional[Any] = preds_index.view(-1 , __lowerCamelCase )[:, 1:] UpperCamelCase :int = decoder(__lowerCamelCase ) UpperCamelCase , UpperCamelCase :Optional[int] = torch.nn.functional.softmax(__lowerCamelCase , dim=2 ).max(dim=2 ) UpperCamelCase :Tuple = preds_max_prob[:, 1:] for index in range(__lowerCamelCase ): UpperCamelCase :Tuple = preds_str[index].find(__lowerCamelCase ) UpperCamelCase :List[Any] = preds_str[index][:pred_eos] UpperCamelCase :List[Any] = preds_index[index].cpu().tolist() UpperCamelCase :Optional[Any] = pred_index.index(__lowerCamelCase ) if eos_token in pred_index else -1 UpperCamelCase :List[str] = preds_max_prob[index][: pred_eos_index + 1] UpperCamelCase :List[str] = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(__lowerCamelCase ) conf_scores.append(__lowerCamelCase ) return dec_strs, conf_scores def _A ( self : Optional[Any] , __lowerCamelCase : str ): UpperCamelCase :Dict = [seq.replace(""" """ , """""" ) for seq in self.char_tokenizer.batch_decode(__lowerCamelCase )] return decode_strs def _A ( self : Union[str, Any] , __lowerCamelCase : str ): return self.bpe_tokenizer.batch_decode(__lowerCamelCase ) def _A ( self : int , __lowerCamelCase : Optional[int] ): UpperCamelCase :Any = [seq.replace(""" """ , """""" ) for seq in self.wp_tokenizer.batch_decode(__lowerCamelCase )] return decode_strs
62
0
import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class UpperCamelCase__ (A__ ): '''simple docstring''' def __get__( self , UpperCamelCase__ , UpperCamelCase__=None ) -> Any: if obj is None: return self if self.fget is None: raise AttributeError("unreadable attribute" ) lowerCamelCase : int = """__cached_""" + self.fget.__name__ lowerCamelCase : Optional[int] = getattr(__lowercase , __lowercase , __lowercase ) if cached is None: lowerCamelCase : int = self.fget(__lowercase ) setattr(__lowercase , __lowercase , __lowercase ) return cached def A ( _SCREAMING_SNAKE_CASE ) -> int: lowerCamelCase : List[Any] = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def A ( _SCREAMING_SNAKE_CASE ) -> Optional[Any]: if is_torch_fx_proxy(_lowercase ): return True if is_torch_available(): import torch if isinstance(_lowercase ,torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(_lowercase ,tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(_lowercase ,(jnp.ndarray, Tracer) ): return True return isinstance(_lowercase ,np.ndarray ) def A ( _SCREAMING_SNAKE_CASE ) -> Dict: return isinstance(_lowercase ,np.ndarray ) def A ( _SCREAMING_SNAKE_CASE ) -> int: return _is_numpy(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> List[str]: import torch return isinstance(_lowercase ,torch.Tensor ) def A ( _SCREAMING_SNAKE_CASE ) -> Any: return False if not is_torch_available() else _is_torch(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Tuple: import torch return isinstance(_lowercase ,torch.device ) def A ( _SCREAMING_SNAKE_CASE ) -> Any: return False if not is_torch_available() else _is_torch_device(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Tuple: import torch if isinstance(_lowercase ,_lowercase ): if hasattr(_lowercase ,_lowercase ): lowerCamelCase : Union[str, Any] = getattr(_lowercase ,_lowercase ) else: return False return isinstance(_lowercase ,torch.dtype ) def A ( _SCREAMING_SNAKE_CASE ) -> Any: return False if not is_torch_available() else _is_torch_dtype(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Optional[Any]: import tensorflow as tf return isinstance(_lowercase ,tf.Tensor ) def A ( _SCREAMING_SNAKE_CASE ) -> Optional[int]: return False if not is_tf_available() else _is_tensorflow(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Optional[Any]: import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(_lowercase ,"is_symbolic_tensor" ): return tf.is_symbolic_tensor(_lowercase ) return type(_lowercase ) == tf.Tensor def A ( _SCREAMING_SNAKE_CASE ) -> Dict: return False if not is_tf_available() else _is_tf_symbolic_tensor(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: import jax.numpy as jnp # noqa: F811 return isinstance(_lowercase ,jnp.ndarray ) def A ( _SCREAMING_SNAKE_CASE ) -> Any: return False if not is_flax_available() else _is_jax(_lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> int: if isinstance(_lowercase ,(dict, UserDict) ): return {k: to_py_obj(_lowercase ) for k, v in obj.items()} elif isinstance(_lowercase ,(list, tuple) ): return [to_py_obj(_lowercase ) for o in obj] elif is_tf_tensor(_lowercase ): return obj.numpy().tolist() elif is_torch_tensor(_lowercase ): return obj.detach().cpu().tolist() elif is_jax_tensor(_lowercase ): return np.asarray(_lowercase ).tolist() elif isinstance(_lowercase ,(np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def A ( _SCREAMING_SNAKE_CASE ) -> Optional[Any]: if isinstance(_lowercase ,(dict, UserDict) ): return {k: to_numpy(_lowercase ) for k, v in obj.items()} elif isinstance(_lowercase ,(list, tuple) ): return np.array(_lowercase ) elif is_tf_tensor(_lowercase ): return obj.numpy() elif is_torch_tensor(_lowercase ): return obj.detach().cpu().numpy() elif is_jax_tensor(_lowercase ): return np.asarray(_lowercase ) else: return obj class UpperCamelCase__ (A__ ): '''simple docstring''' def _lowercase ( self ) -> List[Any]: lowerCamelCase : Union[str, Any] = fields(self ) # Safety and consistency checks if not len(__lowercase ): raise ValueError(F'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(F'''{self.__class__.__name__} should not have more than one required field.''' ) lowerCamelCase : Dict = getattr(self , class_fields[0].name ) lowerCamelCase : List[Any] = all(getattr(self , field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(__lowercase ): if isinstance(__lowercase , __lowercase ): lowerCamelCase : Tuple = first_field.items() lowerCamelCase : Optional[Any] = True else: try: lowerCamelCase : Optional[Any] = iter(__lowercase ) lowerCamelCase : Any = True except TypeError: lowerCamelCase : str = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(__lowercase ): if ( not isinstance(__lowercase , (list, tuple) ) or not len(__lowercase ) == 2 or not isinstance(element[0] , __lowercase ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute lowerCamelCase : Dict = first_field else: # If we have a mixed iterator, raise an error raise ValueError( F'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self , element[0] , element[1] ) if element[1] is not None: lowerCamelCase : Optional[Any] = element[1] elif first_field is not None: lowerCamelCase : List[str] = first_field else: for field in class_fields: lowerCamelCase : List[Any] = getattr(self , field.name ) if v is not None: lowerCamelCase : Optional[int] = v def __delitem__( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> Dict: raise Exception(F'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def _lowercase ( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> Union[str, Any]: raise Exception(F'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def _lowercase ( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> List[Any]: raise Exception(F'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def _lowercase ( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> str: raise Exception(F'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self , UpperCamelCase__ ) -> Optional[int]: if isinstance(__lowercase , __lowercase ): lowerCamelCase : str = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self , UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(__lowercase , __lowercase ) super().__setattr__(__lowercase , __lowercase ) def __setitem__( self , UpperCamelCase__ , UpperCamelCase__ ) -> Tuple: super().__setitem__(__lowercase , __lowercase ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(__lowercase , __lowercase ) def _lowercase ( self ) -> Tuple[Any]: return tuple(self[k] for k in self.keys() ) class UpperCamelCase__ (A__ , A__ ): '''simple docstring''' @classmethod def _lowercase ( cls , UpperCamelCase__ ) -> Tuple: raise ValueError( F'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class UpperCamelCase__ (A__ ): '''simple docstring''' lowerCamelCase_ : str = "longest" lowerCamelCase_ : Tuple = "max_length" lowerCamelCase_ : Optional[int] = "do_not_pad" class UpperCamelCase__ (A__ ): '''simple docstring''' lowerCamelCase_ : List[str] = "pt" lowerCamelCase_ : List[Any] = "tf" lowerCamelCase_ : str = "np" lowerCamelCase_ : int = "jax" class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : str = context_managers lowerCamelCase : Optional[Any] = ExitStack() def __enter__( self ) -> List[str]: for context_manager in self.context_managers: self.stack.enter_context(__lowercase ) def __exit__( self , *UpperCamelCase__ , **UpperCamelCase__ ) -> Dict: self.stack.__exit__(*__lowercase , **__lowercase ) def A ( _SCREAMING_SNAKE_CASE ) -> Dict: lowerCamelCase : Union[str, Any] = infer_framework(_lowercase ) if framework == "tf": lowerCamelCase : int = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": lowerCamelCase : Dict = inspect.signature(model_class.forward ) # PyTorch models else: lowerCamelCase : Tuple = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def A ( _SCREAMING_SNAKE_CASE ) -> List[str]: lowerCamelCase : List[Any] = model_class.__name__ lowerCamelCase : List[Any] = infer_framework(_lowercase ) if framework == "tf": lowerCamelCase : Union[str, Any] = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": lowerCamelCase : List[Any] = inspect.signature(model_class.forward ) # PyTorch models else: lowerCamelCase : Optional[Any] = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = "" ,_SCREAMING_SNAKE_CASE = "." ) -> List[Any]: def _flatten_dict(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE="" ,_SCREAMING_SNAKE_CASE="." ): for k, v in d.items(): lowerCamelCase : List[str] = str(_lowercase ) + delimiter + str(_lowercase ) if parent_key else k if v and isinstance(_lowercase ,_lowercase ): yield from flatten_dict(_lowercase ,_lowercase ,delimiter=_lowercase ).items() else: yield key, v return dict(_flatten_dict(_lowercase ,_lowercase ,_lowercase ) ) @contextmanager def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = False ) -> str: if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=None ) -> Optional[Any]: if is_numpy_array(_lowercase ): return np.transpose(_lowercase ,axes=_lowercase ) elif is_torch_tensor(_lowercase ): return array.T if axes is None else array.permute(*_lowercase ) elif is_tf_tensor(_lowercase ): import tensorflow as tf return tf.transpose(_lowercase ,perm=_lowercase ) elif is_jax_tensor(_lowercase ): return jnp.transpose(_lowercase ,axes=_lowercase ) else: raise ValueError(f'''Type not supported for transpose: {type(_lowercase )}.''' ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> List[Any]: if is_numpy_array(_lowercase ): return np.reshape(_lowercase ,_lowercase ) elif is_torch_tensor(_lowercase ): return array.reshape(*_lowercase ) elif is_tf_tensor(_lowercase ): import tensorflow as tf return tf.reshape(_lowercase ,_lowercase ) elif is_jax_tensor(_lowercase ): return jnp.reshape(_lowercase ,_lowercase ) else: raise ValueError(f'''Type not supported for reshape: {type(_lowercase )}.''' ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=None ) -> Any: if is_numpy_array(_lowercase ): return np.squeeze(_lowercase ,axis=_lowercase ) elif is_torch_tensor(_lowercase ): return array.squeeze() if axis is None else array.squeeze(dim=_lowercase ) elif is_tf_tensor(_lowercase ): import tensorflow as tf return tf.squeeze(_lowercase ,axis=_lowercase ) elif is_jax_tensor(_lowercase ): return jnp.squeeze(_lowercase ,axis=_lowercase ) else: raise ValueError(f'''Type not supported for squeeze: {type(_lowercase )}.''' ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: if is_numpy_array(_lowercase ): return np.expand_dims(_lowercase ,_lowercase ) elif is_torch_tensor(_lowercase ): return array.unsqueeze(dim=_lowercase ) elif is_tf_tensor(_lowercase ): import tensorflow as tf return tf.expand_dims(_lowercase ,axis=_lowercase ) elif is_jax_tensor(_lowercase ): return jnp.expand_dims(_lowercase ,axis=_lowercase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(_lowercase )}.''' ) def A ( _SCREAMING_SNAKE_CASE ) -> int: if is_numpy_array(_lowercase ): return np.size(_lowercase ) elif is_torch_tensor(_lowercase ): return array.numel() elif is_tf_tensor(_lowercase ): import tensorflow as tf return tf.size(_lowercase ) elif is_jax_tensor(_lowercase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(_lowercase )}.''' ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> List[str]: for key, value in auto_map.items(): if isinstance(_lowercase ,(tuple, list) ): lowerCamelCase : Dict = [f'''{repo_id}--{v}''' if (v is not None and """--""" not in v) else v for v in value] elif value is not None and "--" not in value: lowerCamelCase : int = f'''{repo_id}--{value}''' return auto_map def A ( _SCREAMING_SNAKE_CASE ) -> Dict: for base_class in inspect.getmro(_lowercase ): lowerCamelCase : List[str] = base_class.__module__ lowerCamelCase : str = base_class.__name__ if module.startswith("tensorflow" ) or module.startswith("keras" ) or name == "TFPreTrainedModel": return "tf" elif module.startswith("torch" ) or name == "PreTrainedModel": return "pt" elif module.startswith("flax" ) or module.startswith("jax" ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
48
_lowercase : Optional[Any] =[sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_0000)] def lowerCAmelCase_ ( _lowercase : int) -> int: """simple docstring""" a__ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _lowercase : list[bool | None] =[None] * 1000_0000 _lowercase : Tuple =True _lowercase : int =False def lowerCAmelCase_ ( _lowercase : int) -> bool: """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore a__ : Optional[Any] = chain(next_number(_lowercase)) a__ : Dict = number_chain while number < 1000_0000: a__ : Any = number_chain number *= 10 return number_chain def lowerCAmelCase_ ( _lowercase : int = 1000_0000) -> int: """simple docstring""" for i in range(1 , _lowercase): if CHAINS[i] is None: chain(i + 1) return CHAINS[:number].count(_lowercase) if __name__ == "__main__": import doctest doctest.testmod() print(f'{solution() = }')
170
0
'''simple docstring''' from __future__ import annotations __A =10 def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : Optional[int] = 1 UpperCAmelCase__ : Dict = max(UpperCamelCase__ ) while placement <= max_digit: # declare and initialize empty buckets UpperCAmelCase__ : list[list] = [[] for _ in range(UpperCamelCase__ )] # split list_of_ints between the buckets for i in list_of_ints: UpperCAmelCase__ : Union[str, Any] = int((i / placement) % RADIX ) buckets[tmp].append(UpperCamelCase__ ) # put each buckets' contents into list_of_ints UpperCAmelCase__ : List[str] = 0 for b in range(UpperCamelCase__ ): for i in buckets[b]: UpperCAmelCase__ : Tuple = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
283
'''simple docstring''' import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class _snake_case ( a__ ): lowerCAmelCase :Optional[int] = '''''' lowerCAmelCase :str = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowerCAmelCase :str = None # compression type in fsspec. ex: "gzip" lowerCAmelCase :str = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self , _lowerCamelCase = "" , _lowerCamelCase = None , _lowerCamelCase = None , **_lowerCamelCase): super().__init__(self , **_lowerCamelCase) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCAmelCase__ : Optional[Any] = fsspec.open( _lowerCamelCase , mode="""rb""" , protocol=_lowerCamelCase , compression=self.compression , client_kwargs={ """requote_redirect_url""": False, # see https://github.com/huggingface/datasets/pull/5459 """trust_env""": True, # Enable reading proxy env variables. **(target_options or {}).pop("""client_kwargs""" , {}), # To avoid issues if it was already passed. } , **(target_options or {}) , ) UpperCAmelCase__ : List[Any] = os.path.basename(self.file.path.split("""::""")[0]) UpperCAmelCase__ : Dict = ( self.compressed_name[: self.compressed_name.rindex(""".""")] if """.""" in self.compressed_name else self.compressed_name ) UpperCAmelCase__ : Tuple = None @classmethod def snake_case__ ( cls , _lowerCamelCase): # compressed file paths are always relative to the archive root return super()._strip_protocol(_lowerCamelCase).lstrip("""/""") def snake_case__ ( self): if self.dir_cache is None: UpperCAmelCase__ : Optional[Any] = {**self.file.fs.info(self.file.path), """name""": self.uncompressed_name} UpperCAmelCase__ : Union[str, Any] = {f["""name"""]: f} def snake_case__ ( self , _lowerCamelCase): return self.file.open().read() def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase = "rb" , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase , ): UpperCAmelCase__ : List[str] = self._strip_protocol(_lowerCamelCase) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''') return self.file.open() class _snake_case ( a__ ): lowerCAmelCase :Dict = '''bz2''' lowerCAmelCase :List[str] = '''bz2''' lowerCAmelCase :Dict = '''.bz2''' class _snake_case ( a__ ): lowerCAmelCase :int = '''gzip''' lowerCAmelCase :Tuple = '''gzip''' lowerCAmelCase :str = '''.gz''' class _snake_case ( a__ ): lowerCAmelCase :List[str] = '''lz4''' lowerCAmelCase :Any = '''lz4''' lowerCAmelCase :int = '''.lz4''' class _snake_case ( a__ ): lowerCAmelCase :Union[str, Any] = '''xz''' lowerCAmelCase :int = '''xz''' lowerCAmelCase :List[Any] = '''.xz''' class _snake_case ( a__ ): lowerCAmelCase :Tuple = '''zstd''' lowerCAmelCase :List[str] = '''zstd''' lowerCAmelCase :Union[str, Any] = '''.zst''' def __init__( self , _lowerCamelCase , _lowerCamelCase = "rb" , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = DEFAULT_BLOCK_SIZE , **_lowerCamelCase , ): super().__init__( fo=_lowerCamelCase , mode=_lowerCamelCase , target_protocol=_lowerCamelCase , target_options=_lowerCamelCase , block_size=_lowerCamelCase , **_lowerCamelCase , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCAmelCase__ : Dict = self.file.__enter__ class _snake_case : def __init__( self , _lowerCamelCase): UpperCAmelCase__ : Optional[int] = file_ def __enter__( self): self._file.__enter__() return self def __exit__( self , *_lowerCamelCase , **_lowerCamelCase): self._file.__exit__(*_lowerCamelCase , **_lowerCamelCase) def __iter__( self): return iter(self._file) def snake_case__ ( self): return next(self._file) def __getattr__( self , _lowerCamelCase): return getattr(self._file , _lowerCamelCase) def fixed_enter(*_lowerCamelCase , **_lowerCamelCase): return WrappedFile(_enter(*_lowerCamelCase , **_lowerCamelCase)) UpperCAmelCase__ : List[Any] = fixed_enter
283
1