code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import argparse import os import re lowerCAmelCase_ = '''src/transformers/models/auto''' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict lowerCAmelCase_ = re.compile(r'''[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict''') # re pattern that matches identifiers in mappings lowerCAmelCase_ = re.compile(r'''\s*\(\s*"(\S[^"]+)"''') def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase = False ) -> Tuple: """simple docstring""" with open(_UpperCamelCase , '''r''' , encoding='''utf-8''' ) as f: snake_case_ : Optional[Any] = f.read() snake_case_ : Optional[Any] = content.split('''\n''' ) snake_case_ : Optional[int] = [] snake_case_ : List[Any] = 0 while line_idx < len(_UpperCamelCase ): if _re_intro_mapping.search(lines[line_idx] ) is not None: snake_case_ : Any = len(re.search(R'''^(\s*)\S''' , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(''' ''' * indent + '''(''' ): new_lines.append(lines[line_idx] ) line_idx += 1 snake_case_ : List[Any] = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": snake_case_ : int = line_idx while not lines[line_idx].startswith(''' ''' * indent + ''')''' ): line_idx += 1 blocks.append('''\n'''.join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers snake_case_ : int = sorted(_UpperCamelCase , key=lambda _UpperCamelCase : _re_identifier.search(_UpperCamelCase ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(_UpperCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write('''\n'''.join(_UpperCamelCase ) ) elif "\n".join(_UpperCamelCase ) != content: return True def lowerCamelCase_ ( _UpperCamelCase = False ) -> Tuple: """simple docstring""" snake_case_ : Union[str, Any] = [os.path.join(_UpperCamelCase , _UpperCamelCase ) for f in os.listdir(_UpperCamelCase ) if f.endswith('''.py''' )] snake_case_ : Any = [sort_auto_mapping(_UpperCamelCase , overwrite=_UpperCamelCase ) for fname in fnames] if not overwrite and any(_UpperCamelCase ): snake_case_ : List[str] = [f for f, d in zip(_UpperCamelCase , _UpperCamelCase ) if d] raise ValueError( f'''The following files have auto mappings that need sorting: {", ".join(_UpperCamelCase )}. Run `make style` to fix''' ''' this.''' ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') lowerCAmelCase_ = parser.parse_args() sort_all_auto_mappings(not args.check_only)
60
from __future__ import annotations import math def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : bool , __lowerCamelCase : list[int] , __lowerCamelCase : float ): if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(__lowerCamelCase ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = [90, 23, 6, 33, 21, 65, 123, 34423] __UpperCAmelCase : str = math.log(len(__lowerCamelCase ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = (DPMSolverSDEScheduler,) snake_case__ = 1_0 def a ( self : Dict , **SCREAMING_SNAKE_CASE__ : str ) -> str: lowerCAmelCase__ = { "num_train_timesteps": 1_100, "beta_start": 0.0_001, "beta_end": 0.02, "beta_schedule": "linear", "noise_sampler_seed": 0, } config.update(**SCREAMING_SNAKE_CASE__ ) return config def a ( self : Any ) -> int: for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] ) -> Union[str, Any]: for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=SCREAMING_SNAKE_CASE__ , beta_end=SCREAMING_SNAKE_CASE__ ) def a ( self : Tuple ) -> int: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] ) -> Dict: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=SCREAMING_SNAKE_CASE__ ) def a ( self : List[str] ) -> Optional[Any]: lowerCAmelCase__ = self.scheduler_classes[0] lowerCAmelCase__ = self.get_scheduler_config() lowerCAmelCase__ = scheduler_class(**SCREAMING_SNAKE_CASE__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ = self.dummy_model() lowerCAmelCase__ = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ = sample.to(SCREAMING_SNAKE_CASE__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = output.prev_sample lowerCAmelCase__ = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.47_821_044_921_875 ) < 1e-2 assert abs(result_mean.item() - 0.2_178_705_964_565_277 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_352_111_816_406 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_906_892_299_652 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def a ( self : Optional[Any] ) -> Optional[int]: lowerCAmelCase__ = self.scheduler_classes[0] lowerCAmelCase__ = self.get_scheduler_config(prediction_type="v_prediction" ) lowerCAmelCase__ = scheduler_class(**SCREAMING_SNAKE_CASE__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ = self.dummy_model() lowerCAmelCase__ = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ = sample.to(SCREAMING_SNAKE_CASE__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = output.prev_sample lowerCAmelCase__ = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 124.77_149_200_439_453 ) < 1e-2 assert abs(result_mean.item() - 0.16_226_289_014_816_284 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 128.1_663_360_595_703 ) < 1e-2 assert abs(result_mean.item() - 0.16_688_326_001_167_297 ) < 1e-3 else: assert abs(result_sum.item() - 119.8_487_548_828_125 ) < 1e-2 assert abs(result_mean.item() - 0.1_560_530_662_536_621 ) < 1e-3 def a ( self : Tuple ) -> Optional[int]: lowerCAmelCase__ = self.scheduler_classes[0] lowerCAmelCase__ = self.get_scheduler_config() lowerCAmelCase__ = scheduler_class(**SCREAMING_SNAKE_CASE__ ) scheduler.set_timesteps(self.num_inference_steps , device=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.dummy_model() lowerCAmelCase__ = self.dummy_sample_deter.to(SCREAMING_SNAKE_CASE__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase__ = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = output.prev_sample lowerCAmelCase__ = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.46_957_397_460_938 ) < 1e-2 assert abs(result_mean.item() - 0.21_805_934_607_982_635 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_353_637_695_312 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_908_382_415_771 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def a ( self : List[str] ) -> List[str]: lowerCAmelCase__ = self.scheduler_classes[0] lowerCAmelCase__ = self.get_scheduler_config() lowerCAmelCase__ = scheduler_class(**SCREAMING_SNAKE_CASE__ , use_karras_sigmas=SCREAMING_SNAKE_CASE__ ) scheduler.set_timesteps(self.num_inference_steps , device=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.dummy_model() lowerCAmelCase__ = self.dummy_sample_deter.to(SCREAMING_SNAKE_CASE__ ) * scheduler.init_noise_sigma lowerCAmelCase__ = sample.to(SCREAMING_SNAKE_CASE__ ) for t in scheduler.timesteps: lowerCAmelCase__ = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = output.prev_sample lowerCAmelCase__ = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 176.66_974_135_742_188 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 177.63_653_564_453_125 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 else: assert abs(result_sum.item() - 170.3_135_223_388_672 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2
61
from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class a ( lowercase__ ): """simple docstring""" a : Optional[Any] = 'openai-gpt' a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __lowercase : Tuple=40478 , __lowercase : Tuple=512 , __lowercase : int=768 , __lowercase : Dict=12 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[Any]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Optional[Any]=0.1 , __lowercase : Tuple=0.1 , __lowercase : Dict=1e-5 , __lowercase : Any=0.02 , __lowercase : List[str]="cls_index" , __lowercase : str=True , __lowercase : Dict=None , __lowercase : str=True , __lowercase : List[str]=0.1 , **__lowercase : List[Any] , ) -> List[Any]: __UpperCAmelCase : Optional[int] = vocab_size __UpperCAmelCase : Optional[Any] = n_positions __UpperCAmelCase : Optional[int] = n_embd __UpperCAmelCase : str = n_layer __UpperCAmelCase : Any = n_head __UpperCAmelCase : Tuple = afn __UpperCAmelCase : Any = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : str = attn_pdrop __UpperCAmelCase : str = layer_norm_epsilon __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : Optional[int] = summary_type __UpperCAmelCase : Optional[Any] = summary_use_proj __UpperCAmelCase : List[Any] = summary_activation __UpperCAmelCase : Union[str, Any] = summary_first_dropout __UpperCAmelCase : Dict = summary_proj_to_labels super().__init__(**__lowercase )
63
0
class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Dict ): SCREAMING_SNAKE_CASE : Optional[int] = {} def _A ( self : str ): print(self.vertex ) for i in self.vertex: print(UpperCAmelCase_ , " -> " , " -> ".join([str(UpperCAmelCase_ ) for j in self.vertex[i]] ) ) def _A ( self : Any , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(UpperCAmelCase_ ) else: # else make a new vertex SCREAMING_SNAKE_CASE : List[str] = [to_vertex] def _A ( self : int ): # visited array for storing already visited nodes SCREAMING_SNAKE_CASE : Any = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(UpperCAmelCase_ , UpperCAmelCase_ ) def _A ( self : Dict , UpperCAmelCase_ : int , UpperCAmelCase_ : list ): # mark start vertex as visited SCREAMING_SNAKE_CASE : Any = True print(UpperCAmelCase_ , end=" " ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(UpperCAmelCase_ , UpperCAmelCase_ ) if __name__ == "__main__": snake_case = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
62
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = KandinskyVaaInpaintPipeline a : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] a : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] a : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] a : List[Any] = False @property def UpperCAmelCase ( self : int ) -> Dict: return 32 @property def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: return 32 @property def UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def UpperCAmelCase ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return 100 @property def UpperCAmelCase ( self : Dict ) -> Any: torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } __UpperCAmelCase : int = UNetaDConditionModel(**__lowercase ) return model @property def UpperCAmelCase ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self : Any ) -> List[Any]: __UpperCAmelCase : List[str] = self.dummy_unet __UpperCAmelCase : List[str] = self.dummy_movq __UpperCAmelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowercase , ) __UpperCAmelCase : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCAmelCase ( self : str , __lowercase : Tuple , __lowercase : List[str]=0 ) -> Optional[Any]: __UpperCAmelCase : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(__lowercase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask __UpperCAmelCase : Union[str, Any] = np.ones((64, 64) , dtype=np.floataa ) __UpperCAmelCase : List[str] = 0 if str(__lowercase ).startswith("""mps""" ): __UpperCAmelCase : List[str] = torch.manual_seed(__lowercase ) else: __UpperCAmelCase : Optional[int] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __UpperCAmelCase : Optional[Any] = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = """cpu""" __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**__lowercase ) __UpperCAmelCase : Tuple = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(__lowercase ) ) __UpperCAmelCase : Tuple = output.images __UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] __UpperCAmelCase : str = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __UpperCAmelCase : Optional[Any] = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def UpperCAmelCase ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) __UpperCAmelCase : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) __UpperCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa ) __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = """a hat""" __UpperCAmelCase : str = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) __UpperCAmelCase : Any = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) __UpperCAmelCase : int = pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() __UpperCAmelCase : Optional[int] = pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
63
0
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _lowerCamelCase ( UpperCamelCase_ , unittest.TestCase ): __a = TextToVideoSDPipeline __a = TEXT_TO_IMAGE_PARAMS __a = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __a = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase_ ( self ) -> Union[str, Any]: torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__: List[str]= UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) SCREAMING_SNAKE_CASE__: Optional[int]= DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__: int= AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__: str= CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) SCREAMING_SNAKE_CASE__: Dict= CLIPTextModel(lowerCAmelCase ) SCREAMING_SNAKE_CASE__: int= CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) SCREAMING_SNAKE_CASE__: Tuple= { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase_ ( self , lowerCAmelCase , lowerCAmelCase=0 ) -> Dict: if str(lowerCAmelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE__: Any= torch.manual_seed(lowerCAmelCase ) else: SCREAMING_SNAKE_CASE__: Any= torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Optional[Any]= { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase_ ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__: str= '''cpu''' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__: List[Any]= self.get_dummy_components() SCREAMING_SNAKE_CASE__: List[Any]= TextToVideoSDPipeline(**lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Tuple= sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Dict= self.get_dummy_inputs(lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Any= '''np''' SCREAMING_SNAKE_CASE__: Optional[int]= sd_pipe(**lowerCAmelCase ).frames SCREAMING_SNAKE_CASE__: Union[str, Any]= frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) SCREAMING_SNAKE_CASE__: Any= np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> int: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase_ ( self ) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1e-2 ) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def UpperCamelCase_ ( self ) -> List[str]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def UpperCamelCase_ ( self ) -> Optional[int]: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''' ) def UpperCamelCase_ ( self ) -> Dict: pass def UpperCamelCase_ ( self ) -> Dict: return super().test_progress_bar() @slow @skip_mps class _lowerCamelCase ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Dict: SCREAMING_SNAKE_CASE__: Optional[int]= load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''' ) SCREAMING_SNAKE_CASE__: Any= TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) SCREAMING_SNAKE_CASE__: Union[str, Any]= DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) SCREAMING_SNAKE_CASE__: int= pipe.to('''cuda''' ) SCREAMING_SNAKE_CASE__: List[str]= '''Spiderman is surfing''' SCREAMING_SNAKE_CASE__: Tuple= torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__: Optional[int]= pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type='''pt''' ).frames SCREAMING_SNAKE_CASE__: List[str]= video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def UpperCamelCase_ ( self ) -> List[str]: SCREAMING_SNAKE_CASE__: List[str]= load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''' ) SCREAMING_SNAKE_CASE__: List[Any]= TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) SCREAMING_SNAKE_CASE__: Any= pipe.to('''cuda''' ) SCREAMING_SNAKE_CASE__: Optional[Any]= '''Spiderman is surfing''' SCREAMING_SNAKE_CASE__: Tuple= torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__: Dict= pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type='''pt''' ).frames SCREAMING_SNAKE_CASE__: Dict= video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
64
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a : List[Any] = True except ImportError: a : str = False try: from torch.hub import _get_torch_home a : List[Any] = _get_torch_home() except ImportError: a : int = os.path.expanduser( os.getenv("TORCH_HOME", os.path.join(os.getenv("XDG_CACHE_HOME", "~/.cache"), "torch")) ) a : Optional[Any] = os.path.join(torch_cache_home, "transformers") a : Optional[Any] = "https://cdn.huggingface.co" a : List[str] = "https://s3.amazonaws.com/models.huggingface.co/bert" a : Any = "/".join(str(Path(__file__).resolve()).split("/")[:-1]) a : Optional[int] = os.path.join(PATH, "config.yaml") a : Dict = os.path.join(PATH, "attributes.txt") a : Tuple = os.path.join(PATH, "objects.txt") a : Dict = os.getenv("PYTORCH_PRETRAINED_BERT_CACHE", default_cache_path) a : Dict = os.getenv("PYTORCH_TRANSFORMERS_CACHE", PYTORCH_PRETRAINED_BERT_CACHE) a : Optional[int] = os.getenv("TRANSFORMERS_CACHE", PYTORCH_TRANSFORMERS_CACHE) a : Any = "pytorch_model.bin" a : int = "config.yaml" def lowerCamelCase__ ( __lowerCamelCase : str=OBJECTS , __lowerCamelCase : Union[str, Any]=ATTRIBUTES ): __UpperCAmelCase : Union[str, Any] = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(""",""" )[0].lower().strip() ) __UpperCAmelCase : Dict = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(""",""" )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : List[str] = OrderedDict() with open(__lowerCamelCase , """rb""" ) as f: __UpperCAmelCase : int = pkl.load(__lowerCamelCase )["""model"""] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : List[Any] = ckp.pop(__lowerCamelCase ) if isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : Union[str, Any] = torch.tensor(__lowerCamelCase ) else: assert isinstance(__lowerCamelCase , torch.tensor ), type(__lowerCamelCase ) __UpperCAmelCase : List[str] = v return r class a : """simple docstring""" a : Dict = {} def __init__( self : Dict , __lowercase : dict , __lowercase : str = "root" , __lowercase : Any=0 ) -> Dict: __UpperCAmelCase : List[str] = name __UpperCAmelCase : str = level __UpperCAmelCase : int = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : List[str] = copy.deepcopy(__lowercase ) __UpperCAmelCase : Dict = copy.deepcopy(__lowercase ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : Union[str, Any] = Config(__lowercase , name=__lowercase , level=level + 1 ) __UpperCAmelCase : Union[str, Any] = v setattr(self , __lowercase , __lowercase ) __UpperCAmelCase : Any = d def __repr__( self : Optional[Any] ) -> Optional[int]: return str(list((self._pointer.keys()) ) ) def __setattr__( self : List[str] , __lowercase : List[str] , __lowercase : Tuple ) -> int: __UpperCAmelCase : int = val __UpperCAmelCase : List[str] = val __UpperCAmelCase : Union[str, Any] = key.split(""".""" ) __UpperCAmelCase : List[Any] = len(__lowercase ) - 1 __UpperCAmelCase : List[Any] = self._pointer if len(__lowercase ) > 1: for i, l in enumerate(__lowercase ): if hasattr(self , __lowercase ) and isinstance(getattr(self , __lowercase ) , __lowercase ): setattr(getattr(self , __lowercase ) , """.""".join(levels[i:] ) , __lowercase ) if l == last_level: __UpperCAmelCase : Union[str, Any] = val else: __UpperCAmelCase : Union[str, Any] = pointer[l] def UpperCAmelCase ( self : Tuple ) -> Optional[int]: return self._pointer def UpperCAmelCase ( self : str , __lowercase : Optional[int] , __lowercase : Any ) -> Optional[int]: with open(f"""{file_name}""" , """w""" ) as stream: dump(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] ) -> Any: with open(f"""{file_name}""" , """w""" ) as stream: json.dump(__lowercase , __lowercase ) @staticmethod def UpperCAmelCase ( __lowercase : List[Any] ) -> Optional[Any]: with open(__lowercase ) as stream: __UpperCAmelCase : Any = load(__lowercase , Loader=__lowercase ) return data def __str__( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = """ """ if self._name != "root": __UpperCAmelCase : Optional[Any] = f"""{t * (self._level-1)}{self._name}:\n""" else: __UpperCAmelCase : List[Any] = """""" __UpperCAmelCase : Optional[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__lowercase , __lowercase ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__lowercase ).__name__})\n""" __UpperCAmelCase : int = level return r[:-1] @classmethod def UpperCAmelCase ( cls : List[str] , __lowercase : str , **__lowercase : Any ) -> Any: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) return cls(__lowercase ) @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : str , **__lowercase : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase : int = kwargs.pop("""cache_dir""" , __lowercase ) __UpperCAmelCase : int = kwargs.pop("""force_download""" , __lowercase ) __UpperCAmelCase : str = kwargs.pop("""resume_download""" , __lowercase ) __UpperCAmelCase : Dict = kwargs.pop("""proxies""" , __lowercase ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""local_files_only""" , __lowercase ) if os.path.isdir(__lowercase ): __UpperCAmelCase : List[Any] = os.path.join(__lowercase , __lowercase ) elif os.path.isfile(__lowercase ) or is_remote_url(__lowercase ): __UpperCAmelCase : Tuple = pretrained_model_name_or_path else: __UpperCAmelCase : Optional[int] = hf_bucket_url(__lowercase , filename=__lowercase , use_cdn=__lowercase ) try: # Load from URL or cache if already cached __UpperCAmelCase : Optional[int] = cached_path( __lowercase , cache_dir=__lowercase , force_download=__lowercase , proxies=__lowercase , resume_download=__lowercase , local_files_only=__lowercase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : Optional[int] = Config.load_yaml(__lowercase ) except EnvironmentError: __UpperCAmelCase : str = """Can't load config for""" raise EnvironmentError(__lowercase ) if resolved_config_file == config_file: print("""loading configuration file from path""" ) else: print("""loading configuration file cache""" ) return Config.load_yaml(__lowercase ), kwargs def lowerCamelCase__ ( __lowerCamelCase : Dict ): __UpperCAmelCase : Optional[int] = torch.load("""dump.pt""" , map_location=in_tensor.device ) __UpperCAmelCase : Tuple = in_tensor.numpy() __UpperCAmelCase : Optional[int] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception("""tensors are all good""" ) # Hugging face functions below def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Tuple = urlparse(__lowerCamelCase ) return parsed.scheme in ("http", "https") def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : int=True ): __UpperCAmelCase : int = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : Optional[int] = """/""" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=None , ): __UpperCAmelCase : Optional[int] = """python/{}""".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + "; ".join("""{}/{}""".format(__lowerCamelCase , __lowerCamelCase ) for k, v in user_agent.items() ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + user_agent __UpperCAmelCase : List[str] = {"""user-agent""": ua} if resume_size > 0: __UpperCAmelCase : Union[str, Any] = """bytes=%d-""" % (resume_size,) __UpperCAmelCase : Union[str, Any] = requests.get(__lowerCamelCase , stream=__lowerCamelCase , proxies=__lowerCamelCase , headers=__lowerCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : List[str] = response.headers.get("""Content-Length""" ) __UpperCAmelCase : str = resume_size + int(__lowerCamelCase ) if content_length is not None else None __UpperCAmelCase : List[Any] = tqdm( unit="""B""" , unit_scale=__lowerCamelCase , total=__lowerCamelCase , initial=__lowerCamelCase , desc="""Downloading""" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(__lowerCamelCase ) ) temp_file.write(__lowerCamelCase ) progress.close() def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=10 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=None , __lowerCamelCase : List[str]=False , ): if cache_dir is None: __UpperCAmelCase : Optional[Any] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[str] = str(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) __UpperCAmelCase : List[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(__lowerCamelCase , allow_redirects=__lowerCamelCase , proxies=__lowerCamelCase , timeout=__lowerCamelCase ) if response.status_code == 200: __UpperCAmelCase : Dict = response.headers.get("""ETag""" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : List[str] = url_to_filename(__lowerCamelCase , __lowerCamelCase ) # get cache path to put the file __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , __lowerCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(__lowerCamelCase ): return cache_path else: __UpperCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(__lowerCamelCase ) , filename + """.*""" ) if not file.endswith(""".json""" ) and not file.endswith(""".lock""" ) ] if len(__lowerCamelCase ) > 0: return os.path.join(__lowerCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( """Cannot find the requested files in the cached path and outgoing traffic has been""" """ disabled. To enable model look-ups and downloads online, set 'local_files_only'""" """ to False.""" ) return None # From now on, etag is not None. if os.path.exists(__lowerCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : str = cache_path + """.lock""" with FileLock(__lowerCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(__lowerCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : int = cache_path + """.incomplete""" @contextmanager def _resumable_file_manager(): with open(__lowerCamelCase , """a+b""" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(__lowerCamelCase ): __UpperCAmelCase : List[Any] = os.stat(__lowerCamelCase ).st_size else: __UpperCAmelCase : List[Any] = 0 else: __UpperCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=__lowerCamelCase , delete=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( """%s not found in cache or force_download set to True, downloading to %s""" , __lowerCamelCase , temp_file.name , ) http_get( __lowerCamelCase , __lowerCamelCase , proxies=__lowerCamelCase , resume_size=__lowerCamelCase , user_agent=__lowerCamelCase , ) os.replace(temp_file.name , __lowerCamelCase ) __UpperCAmelCase : Any = {"""url""": url, """etag""": etag} __UpperCAmelCase : Union[str, Any] = cache_path + """.json""" with open(__lowerCamelCase , """w""" ) as meta_file: json.dump(__lowerCamelCase , __lowerCamelCase ) return cache_path def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any]=None ): __UpperCAmelCase : Tuple = url.encode("""utf-8""" ) __UpperCAmelCase : Optional[Any] = shaaaa(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = url_hash.hexdigest() if etag: __UpperCAmelCase : int = etag.encode("""utf-8""" ) __UpperCAmelCase : List[str] = shaaaa(__lowerCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(""".h5""" ): filename += ".h5" return filename def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : int=None , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=False , ): if cache_dir is None: __UpperCAmelCase : List[str] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Any = str(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) if is_remote_url(__lowerCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : Tuple = get_from_cache( __lowerCamelCase , cache_dir=__lowerCamelCase , force_download=__lowerCamelCase , proxies=__lowerCamelCase , resume_download=__lowerCamelCase , user_agent=__lowerCamelCase , local_files_only=__lowerCamelCase , ) elif os.path.exists(__lowerCamelCase ): # File, and it exists. __UpperCAmelCase : Tuple = url_or_filename elif urlparse(__lowerCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("""file {} not found""".format(__lowerCamelCase ) ) else: # Something unknown raise ValueError("""unable to parse {} as a URL or as a local path""".format(__lowerCamelCase ) ) if extract_compressed_file: if not is_zipfile(__lowerCamelCase ) and not tarfile.is_tarfile(__lowerCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : int = os.path.split(__lowerCamelCase ) __UpperCAmelCase : Any = output_file.replace(""".""" , """-""" ) + """-extracted""" __UpperCAmelCase : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isdir(__lowerCamelCase ) and os.listdir(__lowerCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : str = output_path + """.lock""" with FileLock(__lowerCamelCase ): shutil.rmtree(__lowerCamelCase , ignore_errors=__lowerCamelCase ) os.makedirs(__lowerCamelCase ) if is_zipfile(__lowerCamelCase ): with ZipFile(__lowerCamelCase , """r""" ) as zip_file: zip_file.extractall(__lowerCamelCase ) zip_file.close() elif tarfile.is_tarfile(__lowerCamelCase ): __UpperCAmelCase : Any = tarfile.open(__lowerCamelCase ) tar_file.extractall(__lowerCamelCase ) tar_file.close() else: raise EnvironmentError("""Archive format of {} could not be identified""".format(__lowerCamelCase ) ) return output_path_extracted return output_path def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : int="," ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): with open(__lowerCamelCase ) as f: __UpperCAmelCase : List[Any] = eval(f.read() ) else: __UpperCAmelCase : List[str] = requests.get(__lowerCamelCase ) try: __UpperCAmelCase : int = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : str = eval(__lowerCamelCase ) except Exception: __UpperCAmelCase : List[Any] = data.split("""\n""" ) req.close() return data def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : Optional[int] = requests.get(__lowerCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCamelCase__ ( __lowerCamelCase : str ): __UpperCAmelCase : int = url.split("""/""" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(__lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as stream: __UpperCAmelCase : List[str] = pkl.load(__lowerCamelCase ) __UpperCAmelCase : Dict = weights.pop("""model""" ) __UpperCAmelCase : Union[str, Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(__lowerCamelCase ) if "running_var" in k: __UpperCAmelCase : Optional[int] = torch.tensor([0] ) __UpperCAmelCase : Tuple = k.replace("""running_var""" , """num_batches_tracked""" ) __UpperCAmelCase : Any = zero return new def lowerCamelCase__ ( ): print(f"""{os.path.abspath(os.path.join(__lowerCamelCase , os.pardir ) )}/demo.ipynb""" ) def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]="RGB" ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): __UpperCAmelCase : List[str] = cva.imread(__lowerCamelCase ) else: __UpperCAmelCase : int = get_image_from_url(__lowerCamelCase ) assert img is not None, f"""could not connect to: {im}""" __UpperCAmelCase : Any = cva.cvtColor(__lowerCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Optional[int] = img[:, :, ::-1] return img def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int=1 ): return (images[i : i + batch] for i in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ))
63
0
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowercase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): @register_to_config def __init__( self : Tuple ,A : int ,A : int ,A : int ,A : float ,A : int ,A : int ,A : int ,A : int ,A : str ,A : bool = False ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : str = nn.Embedding(A ,A ) UpperCAmelCase__ : List[Any] = nn.Embedding(A ,A ) UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : List[str] = nn.Dropout(p=A ) UpperCAmelCase__ : List[Any] = TaConfig( vocab_size=A ,d_model=A ,num_heads=A ,d_kv=A ,d_ff=A ,dropout_rate=A ,feed_forward_proj=A ,is_decoder=A ,is_encoder_decoder=A ,) UpperCAmelCase__ : Dict = nn.ModuleList() for lyr_num in range(A ): UpperCAmelCase__ : int = TaBlock(A ) self.encoders.append(A ) UpperCAmelCase__ : Optional[Any] = TaLayerNorm(A ) UpperCAmelCase__ : int = nn.Dropout(p=A ) def __lowercase ( self : Union[str, Any] ,A : Union[str, Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.token_embedder(A ) UpperCAmelCase__ : List[Any] = encoder_input_tokens.shape[1] UpperCAmelCase__ : str = torch.arange(A ,device=encoder_input_tokens.device ) x += self.position_encoding(A ) UpperCAmelCase__ : Dict = self.dropout_pre(A ) # inverted the attention mask UpperCAmelCase__ : Any = encoder_input_tokens.size() UpperCAmelCase__ : Any = self.get_extended_attention_mask(A ,A ) for lyr in self.encoders: UpperCAmelCase__ : str = lyr(A ,A )[0] UpperCAmelCase__ : List[str] = self.layer_norm(A ) return self.dropout_post(A ), encoder_inputs_mask
65
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Any=13 , __lowercase : Optional[int]=7 , __lowercase : str=True , __lowercase : Optional[Any]=True , __lowercase : int=True , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : int=5 , __lowercase : Tuple=4 , __lowercase : str=37 , __lowercase : Optional[int]="gelu" , __lowercase : Tuple=0.1 , __lowercase : str=0.1 , __lowercase : Dict=512 , __lowercase : List[Any]=16 , __lowercase : Dict=2 , __lowercase : Union[str, Any]=0.02 , __lowercase : Dict=4 , ) -> int: __UpperCAmelCase : Dict = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : str = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Dict = use_token_type_ids __UpperCAmelCase : Optional[int] = use_labels __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : str = type_sequence_label_size __UpperCAmelCase : List[str] = initializer_range __UpperCAmelCase : Tuple = num_choices def UpperCAmelCase ( self : Dict ) -> Tuple: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : List[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase ( self : Tuple ) -> List[Any]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase ( self : Any ) -> List[str]: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[int] = config_and_inputs __UpperCAmelCase : int = True __UpperCAmelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = True a : List[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : List[str] = FlaxRobertaModelTester(self ) @slow def UpperCAmelCase ( self : str ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase : Union[str, Any] = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowercase ) __UpperCAmelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowercase )
63
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu UpperCamelCase = get_tests_dir() + "/test_data/fsmt/fsmt_val_data.json" with io.open(filename, "r", encoding="utf-8") as f: UpperCamelCase = json.load(f) @require_torch class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self , _lowerCAmelCase ): return FSMTTokenizer.from_pretrained(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = FSMTForConditionalGeneration.from_pretrained(_lowerCAmelCase ).to(_lowerCAmelCase ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['en-ru', 26.0], ['ru-en', 22.0], ['en-de', 22.0], ['de-en', 29.0], ] ) @slow def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality _lowercase : Optional[Any] = F"""facebook/wmt19-{pair}""" _lowercase : Any = self.get_tokenizer(_lowerCAmelCase ) _lowercase : Union[str, Any] = self.get_model(_lowerCAmelCase ) _lowercase : Any = bleu_data[pair]['src'] _lowercase : Tuple = bleu_data[pair]['tgt'] _lowercase : Tuple = tokenizer(_lowerCAmelCase , return_tensors='pt' , truncation=_lowerCAmelCase , padding='longest' ).to(_lowerCAmelCase ) _lowercase : List[Any] = model.generate( input_ids=batch.input_ids , num_beams=8 , ) _lowercase : Union[str, Any] = tokenizer.batch_decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) _lowercase : Optional[Any] = calculate_bleu(_lowerCAmelCase , _lowerCAmelCase ) print(_lowerCAmelCase ) self.assertGreaterEqual(scores['bleu'] , _lowerCAmelCase )
66
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging a : Optional[int] = logging.get_logger(__name__) class a ( lowercase__ ): """simple docstring""" a : Tuple = 'linear' a : int = 'cosine' a : Optional[Any] = 'cosine_with_restarts' a : Dict = 'polynomial' a : Tuple = 'constant' a : Dict = 'constant_with_warmup' a : Any = 'piecewise_constant' def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int = -1 ): return LambdaLR(__lowerCamelCase , lambda __lowerCamelCase : 1 , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : str , __lowerCamelCase : int = -1 ): __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Tuple = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __UpperCAmelCase , __UpperCAmelCase : List[str] = rule_str.split(""":""" ) __UpperCAmelCase : Any = int(__lowerCamelCase ) __UpperCAmelCase : List[str] = float(__lowerCamelCase ) __UpperCAmelCase : int = value __UpperCAmelCase : Any = float(rule_list[-1] ) def create_rules_function(__lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): def rule_func(__lowerCamelCase : int ) -> float: __UpperCAmelCase : Tuple = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=-1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float = 0.5 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Dict ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Tuple = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int = 1 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Union[str, Any] ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Union[str, Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=1E-7 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : int=-1 ): __UpperCAmelCase : Tuple = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(f"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __UpperCAmelCase : Optional[Any] = lr_init - lr_end __UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps __UpperCAmelCase : int = 1 - (current_step - num_warmup_steps) / decay_steps __UpperCAmelCase : Optional[int] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a : int = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCamelCase__ ( __lowerCamelCase : Union[str, SchedulerType] , __lowerCamelCase : Optimizer , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 1 , __lowerCamelCase : float = 1.0 , __lowerCamelCase : int = -1 , ): __UpperCAmelCase : Union[str, Any] = SchedulerType(__lowerCamelCase ) __UpperCAmelCase : int = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
63
0
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A_ ( UpperCAmelCase ): """simple docstring""" def __UpperCAmelCase ( self : Union[str, Any] ,__A : str ) -> List[str]: with open(__A ,encoding='utf-8' ) as input_file: _lowercase = re.compile(r'(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)' ) _lowercase = input_file.read() _lowercase = regexp.search(__A ) return match def __UpperCAmelCase ( self : Tuple ,__A : str ) -> Union[str, Any]: with open(__A ,encoding='utf-8' ) as input_file: _lowercase = re.compile(r'#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()' ,re.DOTALL ) _lowercase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _lowercase = regexp.finditer(__A ) _lowercase = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : Any ) -> Tuple: _lowercase = Path('./datasets' ) _lowercase = list(dataset_paths.absolute().glob('**/*.py' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__A ) ): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""" ) def __UpperCAmelCase ( self : Any ) -> List[Any]: _lowercase = Path('./datasets' ) _lowercase = list(dataset_paths.absolute().glob('**/*.py' ) ) for dataset in dataset_files: if self._no_print_statements(str(__A ) ): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""" )
67
from math import pi, sqrt def lowerCamelCase__ ( __lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCamelCase__ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() a : Optional[int] = 1.0 while num: a : List[str] = float(input("Gamma of: ")) print(f"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
63
0
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A = logging.get_logger(__name__) __A = {"vocab_file": "spiece.model"} __A = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class _A ( UpperCamelCase ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Dict=False , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Union[str, Any]=False , __SCREAMING_SNAKE_CASE : Dict="<s>" , __SCREAMING_SNAKE_CASE : Optional[Any]="</s>" , __SCREAMING_SNAKE_CASE : str="<unk>" , __SCREAMING_SNAKE_CASE : str="<sep>" , __SCREAMING_SNAKE_CASE : Any="<pad>" , __SCREAMING_SNAKE_CASE : Tuple="<cls>" , __SCREAMING_SNAKE_CASE : Optional[int]="<mask>" , __SCREAMING_SNAKE_CASE : Tuple=["<eop>", "<eod>"] , __SCREAMING_SNAKE_CASE : Optional[Dict[str, Any]] = None , **__SCREAMING_SNAKE_CASE : Optional[int] , ) -> None: __UpperCAmelCase =AddedToken(__SCREAMING_SNAKE_CASE , lstrip=__SCREAMING_SNAKE_CASE , rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else mask_token __UpperCAmelCase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , additional_special_tokens=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) __UpperCAmelCase =3 __UpperCAmelCase =do_lower_case __UpperCAmelCase =remove_space __UpperCAmelCase =keep_accents __UpperCAmelCase =vocab_file __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) __UpperCAmelCase =jieba __UpperCAmelCase =str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _a ( self : Optional[Any] ) -> List[Any]: return len(self.sp_model ) def _a ( self : Optional[int] ) -> List[Any]: __UpperCAmelCase ={self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ) -> List[str]: __UpperCAmelCase =self.__dict__.copy() __UpperCAmelCase =None return state def __setstate__( self : str , __SCREAMING_SNAKE_CASE : Tuple ) -> str: __UpperCAmelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __UpperCAmelCase ={} __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : Any ) -> Tuple: if self.remove_space: __UpperCAmelCase =""" """.join(inputs.strip().split() ) else: __UpperCAmelCase =inputs __UpperCAmelCase =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: __UpperCAmelCase =unicodedata.normalize("""NFKD""" , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""""".join([c for c in outputs if not unicodedata.combining(__SCREAMING_SNAKE_CASE )] ) if self.do_lower_case: __UpperCAmelCase =outputs.lower() return outputs def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str ) -> List[str]: __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =[] for piece in pieces: if len(__SCREAMING_SNAKE_CASE ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): __UpperCAmelCase =self.sp_model.EncodeAsPieces(piece[:-1].replace(__SCREAMING_SNAKE_CASE , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __UpperCAmelCase =cur_pieces[1:] else: __UpperCAmelCase =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__SCREAMING_SNAKE_CASE ) else: new_pieces.append(__SCREAMING_SNAKE_CASE ) return new_pieces def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> str: return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def _a ( self : str , __SCREAMING_SNAKE_CASE : int ) -> Any: return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Tuple ) -> List[Any]: __UpperCAmelCase ="""""".join(__SCREAMING_SNAKE_CASE ).replace(__SCREAMING_SNAKE_CASE , """ """ ).strip() return out_string def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None ) -> List[int]: __UpperCAmelCase =[self.sep_token_id] __UpperCAmelCase =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _a ( self : str , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None , __SCREAMING_SNAKE_CASE : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__SCREAMING_SNAKE_CASE , token_ids_a=__SCREAMING_SNAKE_CASE , already_has_special_tokens=__SCREAMING_SNAKE_CASE ) if token_ids_a is not None: return ([0] * len(__SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(__SCREAMING_SNAKE_CASE )) + [1, 1] return ([0] * len(__SCREAMING_SNAKE_CASE )) + [1, 1] def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None ) -> List[int]: __UpperCAmelCase =[self.sep_token_id] __UpperCAmelCase =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __UpperCAmelCase =os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE , """wb""" ) as fi: __UpperCAmelCase =self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _a ( self : List[Any] , *__SCREAMING_SNAKE_CASE : Any , **__SCREAMING_SNAKE_CASE : Dict ) -> Dict: __UpperCAmelCase =super()._decode(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
68
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class a : """simple docstring""" a : int a : Node | None = None a : Node | None = None def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = Node(1 ) __UpperCAmelCase : int = Node(2 ) __UpperCAmelCase : Optional[Any] = Node(3 ) __UpperCAmelCase : Dict = Node(4 ) __UpperCAmelCase : Tuple = Node(5 ) return tree def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCamelCase__ ( __lowerCamelCase : Node | None ): __UpperCAmelCase : list[Any] = [] if root is None: return output __UpperCAmelCase : Tuple = deque([root] ) while process_queue: __UpperCAmelCase : Optional[Any] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None ): if root is None: return [] __UpperCAmelCase : list[Sequence[Node | None]] = [] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : int = height(__lowerCamelCase ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : Optional[int] = 0 return output def lowerCamelCase__ ( ): # Main function for testing. __UpperCAmelCase : List[Any] = make_tree() print(f"""In-order Traversal: {inorder(__lowerCamelCase )}""" ) print(f"""Pre-order Traversal: {preorder(__lowerCamelCase )}""" ) print(f"""Post-order Traversal: {postorder(__lowerCamelCase )}""" , """\n""" ) print(f"""Height of Tree: {height(__lowerCamelCase )}""" , """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(__lowerCamelCase ) , """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1 , height(__lowerCamelCase ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(__lowerCamelCase , level=__lowerCamelCase ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
'''simple docstring''' import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) a : int = logging.getLogger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : Tuple ) -> Dict: __snake_case = np.argmax(_UpperCAmelCase , axis=1 ) return np.sum(outputs == labels ) def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Dict: with open(_UpperCAmelCase , encoding="utf_8" ) as f: __snake_case = csv.reader(_UpperCAmelCase ) __snake_case = [] next(_UpperCAmelCase ) # skip the first line for line in tqdm(_UpperCAmelCase ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ) -> List[Any]: __snake_case = [] for dataset in encoded_datasets: __snake_case = len(_UpperCAmelCase ) __snake_case = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __snake_case = np.zeros((n_batch, 2) , dtype=np.intaa ) __snake_case = np.full((n_batch, 2, input_len) , fill_value=-1_00 , dtype=np.intaa ) __snake_case = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCAmelCase ): __snake_case = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __snake_case = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __snake_case = with_conta __snake_case = with_conta __snake_case = len(_UpperCAmelCase ) - 1 __snake_case = len(_UpperCAmelCase ) - 1 __snake_case = with_conta __snake_case = with_conta __snake_case = mc_label __snake_case = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCAmelCase ) for t in all_inputs ) ) return tensor_datasets def __UpperCAmelCase ( ) -> Optional[int]: __snake_case = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_UpperCAmelCase , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=_UpperCAmelCase , type=_UpperCAmelCase , required=_UpperCAmelCase , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=_UpperCAmelCase , default="" ) parser.add_argument("--eval_dataset" , type=_UpperCAmelCase , default="" ) parser.add_argument("--seed" , type=_UpperCAmelCase , default=42 ) parser.add_argument("--num_train_epochs" , type=_UpperCAmelCase , default=3 ) parser.add_argument("--train_batch_size" , type=_UpperCAmelCase , default=8 ) parser.add_argument("--eval_batch_size" , type=_UpperCAmelCase , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=_UpperCAmelCase , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=_UpperCAmelCase , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=_UpperCAmelCase , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=_UpperCAmelCase , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=_UpperCAmelCase , default=6.2_5E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=_UpperCAmelCase , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=_UpperCAmelCase , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=_UpperCAmelCase , default=0.01 ) parser.add_argument("--lm_coef" , type=_UpperCAmelCase , default=0.9 ) parser.add_argument("--n_valid" , type=_UpperCAmelCase , default=3_74 ) parser.add_argument("--server_ip" , type=_UpperCAmelCase , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=_UpperCAmelCase , default="" , help="Can be used for distant debugging." ) __snake_case = parser.parse_args() print(_UpperCAmelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_UpperCAmelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __snake_case = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) __snake_case = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(_UpperCAmelCase , _UpperCAmelCase ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __snake_case = ["_start_", "_delimiter_", "_classify_"] __snake_case = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCAmelCase ) __snake_case = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) __snake_case = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCAmelCase ) ) model.to(_UpperCAmelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCAmelCase : Optional[Any] ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCAmelCase ) ) elif isinstance(_UpperCAmelCase , _UpperCAmelCase ): return obj return [tokenize_and_encode(_UpperCAmelCase ) for o in obj] logger.info("Encoding dataset..." ) __snake_case = load_rocstories_dataset(args.train_dataset ) __snake_case = load_rocstories_dataset(args.eval_dataset ) __snake_case = (train_dataset, eval_dataset) __snake_case = tokenize_and_encode(_UpperCAmelCase ) # Compute the max input length for the Transformer __snake_case = model.config.n_positions // 2 - 2 __snake_case = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __snake_case = min(_UpperCAmelCase , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __snake_case = pre_process_datasets(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , *_UpperCAmelCase ) __snake_case , __snake_case = tensor_datasets[0], tensor_datasets[1] __snake_case = TensorDataset(*_UpperCAmelCase ) __snake_case = RandomSampler(_UpperCAmelCase ) __snake_case = DataLoader(_UpperCAmelCase , sampler=_UpperCAmelCase , batch_size=args.train_batch_size ) __snake_case = TensorDataset(*_UpperCAmelCase ) __snake_case = SequentialSampler(_UpperCAmelCase ) __snake_case = DataLoader(_UpperCAmelCase , sampler=_UpperCAmelCase , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __snake_case = args.max_steps __snake_case = args.max_steps // (len(_UpperCAmelCase ) // args.gradient_accumulation_steps) + 1 else: __snake_case = len(_UpperCAmelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __snake_case = list(model.named_parameters() ) __snake_case = ["bias", "LayerNorm.bias", "LayerNorm.weight"] __snake_case = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] __snake_case = AdamW(_UpperCAmelCase , lr=args.learning_rate , eps=args.adam_epsilon ) __snake_case = get_linear_schedule_with_warmup( _UpperCAmelCase , num_warmup_steps=args.warmup_steps , num_training_steps=_UpperCAmelCase ) if args.do_train: __snake_case , __snake_case , __snake_case = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): __snake_case = 0 __snake_case = 0 __snake_case = tqdm(_UpperCAmelCase , desc="Training" ) for step, batch in enumerate(_UpperCAmelCase ): __snake_case = tuple(t.to(_UpperCAmelCase ) for t in batch ) __snake_case , __snake_case , __snake_case , __snake_case = batch __snake_case = model(_UpperCAmelCase , mc_token_ids=_UpperCAmelCase , lm_labels=_UpperCAmelCase , mc_labels=_UpperCAmelCase ) __snake_case = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __snake_case = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __snake_case = "Training loss: {:.2e} lr: {:.2e}".format(_UpperCAmelCase , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __snake_case = model.module if hasattr(_UpperCAmelCase , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __snake_case = os.path.join(args.output_dir , _UpperCAmelCase ) __snake_case = os.path.join(args.output_dir , _UpperCAmelCase ) torch.save(model_to_save.state_dict() , _UpperCAmelCase ) model_to_save.config.to_json_file(_UpperCAmelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __snake_case = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __snake_case = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCAmelCase ) if args.do_eval: model.eval() __snake_case , __snake_case = 0, 0 __snake_case , __snake_case = 0, 0 for batch in tqdm(_UpperCAmelCase , desc="Evaluating" ): __snake_case = tuple(t.to(_UpperCAmelCase ) for t in batch ) __snake_case , __snake_case , __snake_case , __snake_case = batch with torch.no_grad(): __snake_case , __snake_case , __snake_case , __snake_case = model( _UpperCAmelCase , mc_token_ids=_UpperCAmelCase , lm_labels=_UpperCAmelCase , mc_labels=_UpperCAmelCase ) __snake_case = mc_logits.detach().cpu().numpy() __snake_case = mc_labels.to("cpu" ).numpy() __snake_case = accuracy(_UpperCAmelCase , _UpperCAmelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __snake_case = eval_loss / nb_eval_steps __snake_case = eval_accuracy / nb_eval_examples __snake_case = tr_loss / nb_tr_steps if args.do_train else None __snake_case = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} __snake_case = os.path.join(args.output_dir , "eval_results.txt" ) with open(_UpperCAmelCase , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , _UpperCAmelCase , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
69
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[int] = GPTSanJapaneseTokenizer a : Optional[Any] = False a : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off __UpperCAmelCase : Tuple = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __UpperCAmelCase : Dict = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __UpperCAmelCase : Dict = {"""unk_token""": """<unk>"""} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(__lowercase ) ) def UpperCAmelCase ( self : Tuple , **__lowercase : int ) -> Any: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def UpperCAmelCase ( self : str , __lowercase : Union[str, Any] ) -> Any: __UpperCAmelCase : Any = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __UpperCAmelCase : int = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase ( self : List[Any] , __lowercase : Optional[int] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase : int = self.get_input_output_texts(__lowercase ) __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : Dict = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) return text, ids def UpperCAmelCase ( self : int ) -> Optional[Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Dict ) -> Tuple: pass # TODO add if relevant def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : List[str] = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。 こんばんは、㔺界。""" __UpperCAmelCase : Dict = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids without special tokens __UpperCAmelCase : List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids with special tokens __UpperCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __UpperCAmelCase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : int = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : Tuple = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __UpperCAmelCase : int = """こんにちは、、、、世界。こんばんは、、、、世界。""" __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase ) __UpperCAmelCase : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : int ) -> Optional[int]: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : List[Any] = """こんにちは、世界。""" __UpperCAmelCase : Optional[int] = """こんばんは、㔺界。😀""" __UpperCAmelCase : List[Any] = """こんにちは、世界。こんばんは、世界。😀""" __UpperCAmelCase : List[str] = tokenizer.encode(prefix_text + input_text ) __UpperCAmelCase : List[Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __UpperCAmelCase : Any = tokenizer.encode(__lowercase , prefix_text=__lowercase ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowercase ) __UpperCAmelCase : Any = tokenizer.decode(__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。""" __UpperCAmelCase : List[Any] = """こんばんは、㔺界。😀""" __UpperCAmelCase : Union[str, Any] = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : int = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : List[Any] = [1] + [0] * (len_prefix + len_text + 1) __UpperCAmelCase : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] __UpperCAmelCase : List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCAmelCase : Union[str, Any] = tokenizer(prefix_text + input_text ).token_type_ids __UpperCAmelCase : Optional[Any] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __UpperCAmelCase : Tuple = tokenizer(__lowercase , prefix_text=__lowercase ).token_type_ids self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : List[str] ) -> int: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""あンいワ""" ) __UpperCAmelCase : Tuple = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertNotEqual(__lowercase , __lowercase ) self.assertNotEqual(__lowercase , __lowercase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self : List[Any] ) -> List[str]: __UpperCAmelCase : Any = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : List[Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __UpperCAmelCase : int = tokenizer(__lowercase , padding=__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.batch_encode_plus(__lowercase , padding=__lowercase ) # fmt: off __UpperCAmelCase : Optional[int] = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCAmelCase : Tuple = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCAmelCase : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __lowercase ) self.assertListEqual(x_token.token_type_ids , __lowercase ) self.assertListEqual(x_token.attention_mask , __lowercase ) self.assertListEqual(x_token_a.input_ids , __lowercase ) self.assertListEqual(x_token_a.token_type_ids , __lowercase ) self.assertListEqual(x_token_a.attention_mask , __lowercase ) def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase ( self : Any ) -> int: # tokenizer has no padding token pass
63
0
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = GPTSanJapaneseTokenizer UpperCamelCase = False UpperCamelCase = {'''do_clean_text''': False, '''add_prefix_space''': False} def a__ ( self : Dict ) -> int: """simple docstring""" super().setUp() # fmt: off lowerCamelCase_ = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on lowerCamelCase_ = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 lowerCamelCase_ = {'unk_token': '<unk>'} lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(A_ ) ) def a__ ( self : Any , **A_ : Any ) -> int: """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : Optional[Any] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = 'こんにちは、世界。 \nこんばんは、㔺界。😀' lowerCamelCase_ = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def a__ ( self : List[Any] , A_ : Tuple ) -> str: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.get_input_output_texts(A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.decode(A_ , clean_up_tokenization_spaces=A_ ) return text, ids def a__ ( self : Optional[int] ) -> List[str]: """simple docstring""" pass # TODO add if relevant def a__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" pass # TODO add if relevant def a__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" pass # TODO add if relevant def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() # Testing tokenization lowerCamelCase_ = 'こんにちは、世界。 こんばんは、㔺界。' lowerCamelCase_ = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] lowerCamelCase_ = tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids without special tokens lowerCamelCase_ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] lowerCamelCase_ = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids with special tokens lowerCamelCase_ = tokens + [tokenizer.unk_token] lowerCamelCase_ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] lowerCamelCase_ = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : str ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() # Testing tokenization lowerCamelCase_ = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' lowerCamelCase_ = 'こんにちは、、、、世界。こんばんは、、、、世界。' lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = tokenizer.decode(A_ ) self.assertEqual(A_ , A_ ) @slow def a__ ( self : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization lowerCamelCase_ = 'こんにちは、世界。' lowerCamelCase_ = 'こんばんは、㔺界。😀' lowerCamelCase_ = 'こんにちは、世界。こんばんは、世界。😀' lowerCamelCase_ = tokenizer.encode(prefix_text + input_text ) lowerCamelCase_ = tokenizer.encode('' , prefix_text=prefix_text + input_text ) lowerCamelCase_ = tokenizer.encode(A_ , prefix_text=A_ ) lowerCamelCase_ = tokenizer.decode(A_ ) lowerCamelCase_ = tokenizer.decode(A_ ) lowerCamelCase_ = tokenizer.decode(A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) @slow def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization lowerCamelCase_ = 'こんにちは、世界。' lowerCamelCase_ = 'こんばんは、㔺界。😀' lowerCamelCase_ = len(tokenizer.encode(A_ ) ) - 2 lowerCamelCase_ = len(tokenizer.encode(A_ ) ) - 2 lowerCamelCase_ = [1] + [0] * (len_prefix + len_text + 1) lowerCamelCase_ = [1] * (len_prefix + len_text + 1) + [0] lowerCamelCase_ = [1] + [1] * (len_prefix) + [0] * (len_text + 1) lowerCamelCase_ = tokenizer(prefix_text + input_text ).token_type_ids lowerCamelCase_ = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids lowerCamelCase_ = tokenizer(A_ , prefix_text=A_ ).token_type_ids self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) @slow def a__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) lowerCamelCase_ = tokenizer.encode('あンいワ' ) lowerCamelCase_ = tokenizer.encode('' , prefix_text='あンいワ' ) lowerCamelCase_ = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(A_ ) , tokenizer.decode(A_ ) ) self.assertEqual(tokenizer.decode(A_ ) , tokenizer.decode(A_ ) ) self.assertNotEqual(A_ , A_ ) self.assertNotEqual(A_ , A_ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def a__ ( self : Dict ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) lowerCamelCase_ = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] lowerCamelCase_ = tokenizer(A_ , padding=A_ ) lowerCamelCase_ = tokenizer.batch_encode_plus(A_ , padding=A_ ) # fmt: off lowerCamelCase_ = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] lowerCamelCase_ = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] lowerCamelCase_ = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , A_ ) self.assertListEqual(x_token.token_type_ids , A_ ) self.assertListEqual(x_token.attention_mask , A_ ) self.assertListEqual(x_token_a.input_ids , A_ ) self.assertListEqual(x_token_a.token_type_ids , A_ ) self.assertListEqual(x_token_a.attention_mask , A_ ) def a__ ( self : int ) -> List[str]: """simple docstring""" pass def a__ ( self : Union[str, Any] ) -> Any: """simple docstring""" pass
70
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a : Dict = logging.get_logger(__name__) @dataclass class a ( lowercase__ ): """simple docstring""" a : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : List[Any] , **__lowercase : Dict ) -> Tuple: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase : List[Any] = deprecated_arg[3:] setattr(self , __lowercase , not kwargs.pop(__lowercase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __UpperCAmelCase : str = kwargs.pop("""torchscript""" , self.torchscript ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) __UpperCAmelCase : Optional[Any] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**__lowercase ) a : bool = field(default=lowercase__ , metadata={'help': 'Trace the models using torchscript'} ) a : bool = field(default=lowercase__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) a : str = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def UpperCAmelCase ( self : Any ) -> Tuple["torch.device", int]: requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: __UpperCAmelCase : str = torch.device("""cpu""" ) __UpperCAmelCase : int = 0 elif is_torch_tpu_available(): __UpperCAmelCase : Tuple = xm.xla_device() __UpperCAmelCase : int = 0 else: __UpperCAmelCase : Dict = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) __UpperCAmelCase : Optional[int] = torch.cuda.device_count() return device, n_gpu @property def UpperCAmelCase ( self : Optional[Any] ) -> str: return is_torch_tpu_available() and self.tpu @property def UpperCAmelCase ( self : List[str] ) -> int: requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCAmelCase ( self : int ) -> "torch.device": requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def UpperCAmelCase ( self : int ) -> List[Any]: requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return self.n_gpu > 0
63
0
'''simple docstring''' import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class _snake_case (__SCREAMING_SNAKE_CASE): __A : Dict ="conditional_detr" __A : Tuple =["past_key_values"] __A : Optional[int] ={ "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self ,_snake_case=True ,_snake_case=None ,_snake_case=3 ,_snake_case=3_00 ,_snake_case=6 ,_snake_case=20_48 ,_snake_case=8 ,_snake_case=6 ,_snake_case=20_48 ,_snake_case=8 ,_snake_case=0.0 ,_snake_case=0.0 ,_snake_case=True ,_snake_case="relu" ,_snake_case=2_56 ,_snake_case=0.1 ,_snake_case=0.0 ,_snake_case=0.0 ,_snake_case=0.02 ,_snake_case=1.0 ,_snake_case=False ,_snake_case="sine" ,_snake_case="resnet50" ,_snake_case=True ,_snake_case=False ,_snake_case=2 ,_snake_case=5 ,_snake_case=2 ,_snake_case=1 ,_snake_case=1 ,_snake_case=2 ,_snake_case=5 ,_snake_case=2 ,_snake_case=0.25 ,**_snake_case ,): if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(_snake_case ,_snake_case ): UpperCAmelCase_ : Dict = backbone_config.get("model_type" ) UpperCAmelCase_ : Any = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : List[str] = config_class.from_dict(_snake_case ) UpperCAmelCase_ : Union[str, Any] = use_timm_backbone UpperCAmelCase_ : List[Any] = backbone_config UpperCAmelCase_ : Dict = num_channels UpperCAmelCase_ : List[str] = num_queries UpperCAmelCase_ : Tuple = d_model UpperCAmelCase_ : Any = encoder_ffn_dim UpperCAmelCase_ : str = encoder_layers UpperCAmelCase_ : str = encoder_attention_heads UpperCAmelCase_ : int = decoder_ffn_dim UpperCAmelCase_ : Union[str, Any] = decoder_layers UpperCAmelCase_ : Any = decoder_attention_heads UpperCAmelCase_ : int = dropout UpperCAmelCase_ : Tuple = attention_dropout UpperCAmelCase_ : List[str] = activation_dropout UpperCAmelCase_ : Optional[int] = activation_function UpperCAmelCase_ : List[Any] = init_std UpperCAmelCase_ : str = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Any = decoder_layerdrop UpperCAmelCase_ : Any = encoder_layers UpperCAmelCase_ : Optional[Any] = auxiliary_loss UpperCAmelCase_ : str = position_embedding_type UpperCAmelCase_ : Optional[int] = backbone UpperCAmelCase_ : Dict = use_pretrained_backbone UpperCAmelCase_ : List[Any] = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : Union[str, Any] = giou_cost # Loss coefficients UpperCAmelCase_ : Dict = mask_loss_coefficient UpperCAmelCase_ : List[Any] = dice_loss_coefficient UpperCAmelCase_ : int = cls_loss_coefficient UpperCAmelCase_ : Any = bbox_loss_coefficient UpperCAmelCase_ : Optional[int] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = focal_alpha super().__init__(is_encoder_decoder=_snake_case ,**_snake_case ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): UpperCAmelCase_ : int = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCAmelCase_ : List[str] = self.backbone_config.to_dict() UpperCAmelCase_ : Union[str, Any] = self.__class__.model_type return output class _snake_case (__SCREAMING_SNAKE_CASE): __A : Optional[Any] =version.parse("1.11") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): return 1E-5 @property def UpperCamelCase__ ( self ): return 12
71
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __UpperCAmelCase : str = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Any ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __UpperCAmelCase : Any = features.copy() if features else default_expected_features __UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=__lowerCamelCase , cache_dir=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: __UpperCAmelCase : Dict = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): __UpperCAmelCase : Optional[int] = tmp_path / """cache""" __UpperCAmelCase : str = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() __UpperCAmelCase : Optional[int] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Dict = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): __UpperCAmelCase : int = tmp_path / """cache""" __UpperCAmelCase : int = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Any = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Optional[int] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() with pytest.raises(__lowerCamelCase ): SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
63
0
'''simple docstring''' import math import os import sys def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' lowercase ='''''' try: with open(lowercase_ , '''rb''' ) as binary_file: lowercase =binary_file.read() for dat in data: lowercase =f'{dat:08b}' result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def UpperCamelCase ( lowercase_ : dict[str, str] , lowercase_ : str , lowercase_ : int , lowercase_ : str ) -> None: '''simple docstring''' lexicon.pop(lowercase_ ) lowercase =last_match_id if math.loga(lowercase_ ).is_integer(): for curr_key in lexicon: lowercase ='''0''' + lexicon[curr_key] lowercase =bin(lowercase_ )[2:] def UpperCamelCase ( lowercase_ : str ) -> str: '''simple docstring''' lowercase ={'''0''': '''0''', '''1''': '''1'''} lowercase , lowercase ='''''', '''''' lowercase =len(lowercase_ ) for i in range(len(lowercase_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue lowercase =lexicon[curr_string] result += last_match_id add_key_to_lexicon(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) index += 1 lowercase ='''''' while curr_string != "" and curr_string not in lexicon: curr_string += "0" if curr_string != "": lowercase =lexicon[curr_string] result += last_match_id return result def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> str: '''simple docstring''' lowercase =os.path.getsize(lowercase_ ) lowercase =bin(lowercase_ )[2:] lowercase =len(lowercase_ ) return "0" * (length_length - 1) + file_length_binary + compressed def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> None: '''simple docstring''' lowercase =8 try: with open(lowercase_ , '''wb''' ) as opened_file: lowercase =[ to_write[i : i + byte_length] for i in range(0 , len(lowercase_ ) , lowercase_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array: opened_file.write(int(lowercase_ , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> None: '''simple docstring''' lowercase =read_file_binary(lowercase_ ) lowercase =compress_data(lowercase_ ) lowercase =add_file_length(lowercase_ , lowercase_ ) write_file_binary(lowercase_ , lowercase_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
72
from __future__ import annotations a : Optional[Any] = [True] * 1_000_001 a : Union[str, Any] = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): a : Optional[Any] = False i += 1 def lowerCamelCase__ ( __lowerCamelCase : int ): return seive[n] def lowerCamelCase__ ( __lowerCamelCase : int ): return any(digit in """02468""" for digit in str(__lowerCamelCase ) ) def lowerCamelCase__ ( __lowerCamelCase : int = 1000000 ): __UpperCAmelCase : Optional[Any] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__lowerCamelCase ) and not contains_an_even_digit(__lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) __UpperCAmelCase : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__lowerCamelCase ) )] if all(is_prime(__lowerCamelCase ) for i in list_nums ): result.append(__lowerCamelCase ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"""{len(find_circular_primes()) = }""")
63
0
import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter a_ : Any = True except ImportError: a_ : Optional[Any] = False a_ : Optional[int] = logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCamelCase__ (_UpperCAmelCase): return AddNewModelCommand(args.testing , args.testing_file , path=args.path) class _snake_case ( A__ ): @staticmethod def SCREAMING_SNAKE_CASE__ ( a) -> int: SCREAMING_SNAKE_CASE = parser.add_parser('add-new-model') add_new_model_parser.add_argument('--testing' , action='store_true' , help='If in testing mode.') add_new_model_parser.add_argument('--testing_file' , type=a , help='Configuration file on which to run.') add_new_model_parser.add_argument( '--path' , type=a , help='Path to cookiecutter. Should only be used for testing purposes.') add_new_model_parser.set_defaults(func=a) def __init__( self , a , a , a=None , *a) -> Union[str, Any]: SCREAMING_SNAKE_CASE = testing SCREAMING_SNAKE_CASE = testing_file SCREAMING_SNAKE_CASE = path def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: warnings.warn( 'The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ' 'It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ' 'checks, you should use `transformers-cli add-new-model-like` instead.') if not _has_cookiecutter: raise ImportError( 'Model creation dependencies are required to use the `add_new_model` command. Install them by running ' 'the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n') # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory SCREAMING_SNAKE_CASE = [directory for directory in os.listdir() if 'cookiecutter-template-' == directory[:22]] if len(a) > 0: raise ValueError( 'Several directories starting with `cookiecutter-template-` in current working directory. ' 'Please clean your directory by removing all folders starting with `cookiecutter-template-` or ' 'change your working directory.') SCREAMING_SNAKE_CASE = ( Path(a).parent.parent.parent.parent if self._path is None else Path(self._path).parent.parent ) SCREAMING_SNAKE_CASE = path_to_transformer_root / 'templates' / 'adding_a_new_model' # Execute cookiecutter if not self._testing: cookiecutter(str(a)) else: with open(self._testing_file , 'r') as configuration_file: SCREAMING_SNAKE_CASE = json.load(a) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path) , no_input=a , extra_context=a , ) SCREAMING_SNAKE_CASE = [directory for directory in os.listdir() if 'cookiecutter-template-' in directory[:22]][0] # Retrieve configuration with open(directory + '/configuration.json' , 'r') as configuration_file: SCREAMING_SNAKE_CASE = json.load(a) SCREAMING_SNAKE_CASE = configuration['lowercase_modelname'] SCREAMING_SNAKE_CASE = configuration['generate_tensorflow_pytorch_and_flax'] os.remove(f'''{directory}/configuration.json''') SCREAMING_SNAKE_CASE = 'PyTorch' in generate_tensorflow_pytorch_and_flax SCREAMING_SNAKE_CASE = 'TensorFlow' in generate_tensorflow_pytorch_and_flax SCREAMING_SNAKE_CASE = 'Flax' in generate_tensorflow_pytorch_and_flax SCREAMING_SNAKE_CASE = f'''{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}''' os.makedirs(a , exist_ok=a) os.makedirs(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}''' , exist_ok=a) # Tests require submodules as they have parent imports with open(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py''' , 'w'): pass shutil.move( f'''{directory}/__init__.py''' , f'''{model_dir}/__init__.py''' , ) shutil.move( f'''{directory}/configuration_{lowercase_model_name}.py''' , f'''{model_dir}/configuration_{lowercase_model_name}.py''' , ) def remove_copy_lines(a): with open(a , 'r') as f: SCREAMING_SNAKE_CASE = f.readlines() with open(a , 'w') as f: for line in lines: if "# Copied from transformers." not in line: f.write(a) if output_pytorch: if not self._testing: remove_copy_lines(f'''{directory}/modeling_{lowercase_model_name}.py''') shutil.move( f'''{directory}/modeling_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_{lowercase_model_name}.py''') os.remove(f'''{directory}/test_modeling_{lowercase_model_name}.py''') if output_tensorflow: if not self._testing: remove_copy_lines(f'''{directory}/modeling_tf_{lowercase_model_name}.py''') shutil.move( f'''{directory}/modeling_tf_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_tf_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_tf_{lowercase_model_name}.py''') os.remove(f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''') if output_flax: if not self._testing: remove_copy_lines(f'''{directory}/modeling_flax_{lowercase_model_name}.py''') shutil.move( f'''{directory}/modeling_flax_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_flax_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_flax_{lowercase_model_name}.py''') os.remove(f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''') shutil.move( f'''{directory}/{lowercase_model_name}.md''' , f'''{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md''' , ) shutil.move( f'''{directory}/tokenization_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/tokenization_fast_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}_fast.py''' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(a , a , a): # Create temp file SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = mkstemp() SCREAMING_SNAKE_CASE = False with fdopen(a , 'w') as new_file: with open(a) as old_file: for line in old_file: new_file.write(a) if line_to_copy_below in line: SCREAMING_SNAKE_CASE = True for line_to_copy in lines_to_copy: new_file.write(a) if not line_found: raise ValueError(f'''Line {line_to_copy_below} was not found in file.''') # Copy the file permissions from the old file to the new file copymode(a , a) # Remove original file remove(a) # Move new file move(a , a) def skip_units(a): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(a): with open(a) as datafile: SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False for line in datafile: if "# To replace in: " in line and "##" not in line: SCREAMING_SNAKE_CASE = line.split('"')[1] SCREAMING_SNAKE_CASE = skip_units(a) elif "# Below: " in line and "##" not in line: SCREAMING_SNAKE_CASE = line.split('"')[1] SCREAMING_SNAKE_CASE = skip_units(a) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(a , a , a) SCREAMING_SNAKE_CASE = [] elif "# Replace with" in line and "##" not in line: SCREAMING_SNAKE_CASE = [] elif "##" not in line: lines_to_copy.append(a) remove(a) replace_in_files(f'''{directory}/to_replace_{lowercase_model_name}.py''') os.rmdir(a)
73
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() a : Dict = logging.get_logger(__name__) a : Tuple = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowerCamelCase__ ( __lowerCamelCase : Dict ): if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase : Union[str, Any] = k.replace(__lowerCamelCase , __lowerCamelCase ) if k.startswith("""encoder""" ): __UpperCAmelCase : List[str] = k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase : Optional[Any] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : Union[str, Any] = k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase : Optional[int] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : List[Any] = k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase : Any = k.replace("""norm3""" , """final_layer_norm""" ) return k def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Optional[Any] = [ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase : Dict = sd.pop(__lowerCamelCase ) __UpperCAmelCase : List[str] = k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase : List[str] = v a : Optional[int] = ["START"] @torch.no_grad() def lowerCamelCase__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : str ): __UpperCAmelCase : str = torch.load(__lowerCamelCase , map_location="""cpu""" ) __UpperCAmelCase : Tuple = model["""model"""] __UpperCAmelCase : int = BlenderbotConfig.from_json_file(__lowerCamelCase ) __UpperCAmelCase : List[str] = BlenderbotForConditionalGeneration(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = m.model.state_dict().keys() __UpperCAmelCase : Any = [] __UpperCAmelCase : Any = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase : int = rename_state_dict_key(__lowerCamelCase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase : str = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__lowerCamelCase ) m.model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) m.half() m.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) a : Any = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
63
0
import argparse import torch from transformers import BertForMaskedLM if __name__ == "__main__": lowercase_ = argparse.ArgumentParser( description=( """Extraction some layers of the full BertForMaskedLM or RObertaForMaskedLM for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""bert""", choices=["""bert"""]) parser.add_argument("""--model_name""", default="""bert-base-uncased""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_bert-base-uncased_0247911.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") lowercase_ = parser.parse_args() if args.model_type == "bert": lowercase_ = BertForMaskedLM.from_pretrained(args.model_name) lowercase_ = """bert""" else: raise ValueError("""args.model_type should be \"bert\".""") lowercase_ = model.state_dict() lowercase_ = {} for w in ["word_embeddings", "position_embeddings"]: lowercase_ = state_dict[f'''{prefix}.embeddings.{w}.weight'''] for w in ["weight", "bias"]: lowercase_ = state_dict[f'''{prefix}.embeddings.LayerNorm.{w}'''] lowercase_ = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: for w in ["weight", "bias"]: lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.attention.self.query.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.attention.self.key.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.attention.self.value.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.attention.output.dense.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.attention.output.LayerNorm.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.intermediate.dense.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.output.dense.{w}''' ] lowercase_ = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.output.LayerNorm.{w}''' ] std_idx += 1 lowercase_ = state_dict["""cls.predictions.decoder.weight"""] lowercase_ = state_dict["""cls.predictions.bias"""] if args.vocab_transform: for w in ["weight", "bias"]: lowercase_ = state_dict[f'''cls.predictions.transform.dense.{w}'''] lowercase_ = state_dict[f'''cls.predictions.transform.LayerNorm.{w}'''] print(f'''N layers selected for distillation: {std_idx}''') print(f'''Number of params transferred for distillation: {len(compressed_sd.keys())}''') print(f'''Save transferred checkpoint to {args.dump_checkpoint}.''') torch.save(compressed_sd, args.dump_checkpoint)
74
def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ): __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : List[str] = len(__lowerCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Union[str, Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None __UpperCAmelCase : str = sorted_collection[point] if current_item == item: return point else: if point < left: __UpperCAmelCase : Optional[Any] = left __UpperCAmelCase : Tuple = point elif point > right: __UpperCAmelCase : Optional[Any] = right __UpperCAmelCase : Dict = point else: if item < current_item: __UpperCAmelCase : Union[str, Any] = point - 1 else: __UpperCAmelCase : str = point + 1 return None def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) elif point > right: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , point - 1 ) else: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , point + 1 , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : int ): if collection != sorted(__lowerCamelCase ): raise ValueError("""Collection must be ascending sorted""" ) return True if __name__ == "__main__": import sys a : Optional[Any] = 0 if debug == 1: a : Optional[Any] = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") a : Tuple = 67 a : List[Any] = interpolation_search(collection, target) if result is not None: print(f"""{target} found at positions: {result}""") else: print("Not found")
63
0
'''simple docstring''' from math import loga def a__ ( lowerCAmelCase__ ) -> int: if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise TypeError('''Input value must be a \'int\' type''' ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
75
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
63
0
"""simple docstring""" from __future__ import annotations a_ = [True] * 1_0_0_0_0_0_1 a_ = 2 while i * i <= 1_0_0_0_0_0_0: if seive[i]: for j in range(i * i, 1_0_0_0_0_0_1, i): a_ = False i += 1 def __UpperCAmelCase ( __UpperCamelCase ): return seive[n] def __UpperCAmelCase ( __UpperCamelCase ): return any(digit in '''02468''' for digit in str(__UpperCamelCase ) ) def __UpperCAmelCase ( __UpperCamelCase = 1_00_00_00 ): __lowercase : int = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__UpperCamelCase ) and not contains_an_even_digit(__UpperCamelCase ): __lowercase : Dict = str(__UpperCamelCase ) __lowercase : Dict = [int(str_num[j:] + str_num[:j] ) for j in range(len(__UpperCamelCase ) )] if all(is_prime(__UpperCamelCase ) for i in list_nums ): result.append(__UpperCamelCase ) return result def __UpperCAmelCase ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(F"{len(find_circular_primes()) = }")
76
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class a ( lowercase__ , lowercase__ ): """simple docstring""" a : Dict = 1 @register_to_config def __init__( self : int , __lowercase : int = 1000 , __lowercase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: # set `betas`, `alphas`, `timesteps` self.set_timesteps(__lowercase ) # standard deviation of the initial noise distribution __UpperCAmelCase : List[Any] = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __UpperCAmelCase : List[Any] = 4 # running values __UpperCAmelCase : str = [] def UpperCAmelCase ( self : Union[str, Any] , __lowercase : int , __lowercase : Union[str, torch.device] = None ) -> int: __UpperCAmelCase : int = num_inference_steps __UpperCAmelCase : Union[str, Any] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __UpperCAmelCase : Union[str, Any] = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __UpperCAmelCase : Dict = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __UpperCAmelCase : Dict = torch.sin(steps * math.pi / 2 ) ** 2 __UpperCAmelCase : List[Any] = (1.0 - self.betas**2) ** 0.5 __UpperCAmelCase : Tuple = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __UpperCAmelCase : Dict = timesteps.to(__lowercase ) __UpperCAmelCase : Optional[Any] = [] def UpperCAmelCase ( self : Optional[int] , __lowercase : torch.FloatTensor , __lowercase : int , __lowercase : torch.FloatTensor , __lowercase : bool = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __UpperCAmelCase : List[str] = (self.timesteps == timestep).nonzero().item() __UpperCAmelCase : Optional[Any] = timestep_index + 1 __UpperCAmelCase : List[str] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowercase ) if len(self.ets ) == 1: __UpperCAmelCase : Tuple = self.ets[-1] elif len(self.ets ) == 2: __UpperCAmelCase : Union[str, Any] = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __UpperCAmelCase : Union[str, Any] = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __UpperCAmelCase : List[Any] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __UpperCAmelCase : Union[str, Any] = self._get_prev_sample(__lowercase , __lowercase , __lowercase , __lowercase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowercase ) def UpperCAmelCase ( self : Optional[Any] , __lowercase : torch.FloatTensor , *__lowercase : Optional[Any] , **__lowercase : Any ) -> torch.FloatTensor: return sample def UpperCAmelCase ( self : Tuple , __lowercase : Tuple , __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Dict ) -> str: __UpperCAmelCase : int = self.alphas[timestep_index] __UpperCAmelCase : Tuple = self.betas[timestep_index] __UpperCAmelCase : Any = self.alphas[prev_timestep_index] __UpperCAmelCase : List[str] = self.betas[prev_timestep_index] __UpperCAmelCase : List[str] = (sample - sigma * ets) / max(__lowercase , 1e-8 ) __UpperCAmelCase : List[Any] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Tuple ) -> str: return self.config.num_train_timesteps
63
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Union[str, Any] , UpperCamelCase_ : Dict , UpperCamelCase_ : int=7 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=18 , UpperCamelCase_ : Union[str, Any]=30 , UpperCamelCase_ : int=400 , UpperCamelCase_ : List[str]=True , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Any=True , UpperCamelCase_ : List[str]=False , UpperCamelCase_ : Dict=True , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : List[Any]=[0.5, 0.5, 0.5] , UpperCamelCase_ : List[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = parent __UpperCAmelCase : Dict = batch_size __UpperCAmelCase : List[Any] = num_channels __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : Optional[Any] = min_resolution __UpperCAmelCase : Optional[Any] = max_resolution __UpperCAmelCase : List[Any] = do_resize __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 20} __UpperCAmelCase : List[Any] = do_thumbnail __UpperCAmelCase : str = do_align_axis __UpperCAmelCase : int = do_pad __UpperCAmelCase : Optional[int] = do_normalize __UpperCAmelCase : List[Any] = image_mean __UpperCAmelCase : int = image_std def a_ ( self : List[str]): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = DonutImageProcessor if is_vision_available() else None def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : str = DonutImageProcessingTester(self) @property def a_ ( self : List[Any]): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) self.assertTrue(hasattr(UpperCamelCase_ , "do_thumbnail")) self.assertTrue(hasattr(UpperCamelCase_ , "do_align_long_axis")) self.assertTrue(hasattr(UpperCamelCase_ , "do_pad")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"height": 18, "width": 20}) __UpperCAmelCase : str = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"height": 42, "width": 42}) # Previous config had dimensions in (width, height) order __UpperCAmelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84)) self.assertEqual(image_processor.size , {"height": 84, "width": 42}) def a_ ( self : Optional[Any]): """simple docstring""" pass @is_flaky() def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[str] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : str = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
77
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCamelCase__ ( ): __UpperCAmelCase : Union[str, Any] = ArgumentParser("""Transformers CLI tool""" , usage="""transformers-cli <command> [<args>]""" ) __UpperCAmelCase : Any = parser.add_subparsers(help="""transformers-cli command helpers""" ) # Register commands ConvertCommand.register_subcommand(__lowerCamelCase ) DownloadCommand.register_subcommand(__lowerCamelCase ) EnvironmentCommand.register_subcommand(__lowerCamelCase ) RunCommand.register_subcommand(__lowerCamelCase ) ServeCommand.register_subcommand(__lowerCamelCase ) UserCommands.register_subcommand(__lowerCamelCase ) AddNewModelCommand.register_subcommand(__lowerCamelCase ) AddNewModelLikeCommand.register_subcommand(__lowerCamelCase ) LfsCommands.register_subcommand(__lowerCamelCase ) PTtoTFCommand.register_subcommand(__lowerCamelCase ) # Let's go __UpperCAmelCase : Optional[Any] = parser.parse_args() if not hasattr(__lowerCamelCase , """func""" ): parser.print_help() exit(1 ) # Run __UpperCAmelCase : Tuple = args.func(__lowerCamelCase ) service.run() if __name__ == "__main__": main()
63
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE_: Optional[Any] ={ 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: Any =['PerceiverFeatureExtractor'] SCREAMING_SNAKE_CASE_: Tuple =['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: int =[ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_: Optional[Any] =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
78
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : List[str] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Dict = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
63
0
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType SCREAMING_SNAKE_CASE__ : Optional[List[str]] = None SCREAMING_SNAKE_CASE__ : Optional[Any] = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image SCREAMING_SNAKE_CASE__ : List[Any] = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class UpperCAmelCase_ : __lowerCamelCase = True __lowerCamelCase = None # Automatically constructed __lowerCamelCase = "PIL.Image.Image" __lowerCamelCase = pa.struct({'bytes': pa.binary(), 'path': pa.string()} ) __lowerCamelCase = field(default='Image' , init=__lowerCamelCase , repr=__lowerCamelCase ) def __call__( self ): return self.pa_type def __UpperCAmelCase ( self , _lowerCAmelCase ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Dict = np.array(_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): return {"path": value, "bytes": None} elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): return {"path": None, "bytes": value} elif isinstance(_lowerCAmelCase , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(_lowerCAmelCase ) elif isinstance(_lowerCAmelCase , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(_lowerCAmelCase ) elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( f"An image sample should have one of 'path' or 'bytes' but they are missing or None in {value}." ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Image(decode=True) instead.""" ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support decoding images, please install 'Pillow'.""" ) if token_per_repo_id is None: UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ , UpperCAmelCase__ : Tuple = value["""path"""], value["""bytes"""] if bytes_ is None: if path is None: raise ValueError(f"An image should have one of 'path' or 'bytes' but both are None in {value}." ) else: if is_local_path(_lowerCAmelCase ): UpperCAmelCase__ : str = PIL.Image.open(_lowerCAmelCase ) else: UpperCAmelCase__ : Any = path.split("""::""" )[-1] try: UpperCAmelCase__ : int = string_to_dict(_lowerCAmelCase , config.HUB_DATASETS_URL )["""repo_id"""] UpperCAmelCase__ : int = token_per_repo_id.get(_lowerCAmelCase ) except ValueError: UpperCAmelCase__ : Any = None with xopen(_lowerCAmelCase , """rb""" , use_auth_token=_lowerCAmelCase ) as f: UpperCAmelCase__ : str = BytesIO(f.read() ) UpperCAmelCase__ : str = PIL.Image.open(bytes_ ) else: UpperCAmelCase__ : List[Any] = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def __UpperCAmelCase ( self ): from .features import Value return ( self if self.decode else { "bytes": Value("""binary""" ), "path": Value("""string""" ), } ) def __UpperCAmelCase ( self , _lowerCAmelCase ): if pa.types.is_string(storage.type ): UpperCAmelCase__ : Any = pa.array([None] * len(_lowerCAmelCase ) , type=pa.binary() ) UpperCAmelCase__ : Union[str, Any] = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCAmelCase__ : Optional[int] = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) UpperCAmelCase__ : int = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: UpperCAmelCase__ : Optional[Any] = storage.field("""bytes""" ) else: UpperCAmelCase__ : int = pa.array([None] * len(_lowerCAmelCase ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: UpperCAmelCase__ : Dict = storage.field("""path""" ) else: UpperCAmelCase__ : Dict = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) UpperCAmelCase__ : Tuple = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): UpperCAmelCase__ : str = pa.array( [encode_np_array(np.array(_lowerCAmelCase ) )["""bytes"""] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) UpperCAmelCase__ : Any = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) UpperCAmelCase__ : Optional[int] = pa.StructArray.from_arrays( [bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(_lowerCAmelCase , self.pa_type ) def __UpperCAmelCase ( self , _lowerCAmelCase ): @no_op_if_value_is_null def path_to_bytes(_lowerCAmelCase ): with xopen(_lowerCAmelCase , """rb""" ) as f: UpperCAmelCase__ : Optional[int] = f.read() return bytes_ UpperCAmelCase__ : Optional[Any] = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) UpperCAmelCase__ : Optional[int] = pa.array( [os.path.basename(_lowerCAmelCase ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) UpperCAmelCase__ : List[Any] = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(_lowerCAmelCase , self.pa_type ) def _lowerCamelCase ( ) -> List[str]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() UpperCAmelCase__ : List[Any] = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _lowerCamelCase ( __lowerCamelCase ) -> bytes: '''simple docstring''' UpperCAmelCase__ : Tuple = BytesIO() if image.format in list_image_compression_formats(): UpperCAmelCase__ : Union[str, Any] = image.format else: UpperCAmelCase__ : List[str] = """PNG""" if image.mode in ["""1""", """L""", """LA""", """RGB""", """RGBA"""] else """TIFF""" image.save(__lowerCamelCase , format=__lowerCamelCase ) return buffer.getvalue() def _lowerCamelCase ( __lowerCamelCase ) -> dict: '''simple docstring''' if hasattr(__lowerCamelCase , """filename""" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__lowerCamelCase )} def _lowerCamelCase ( __lowerCamelCase ) -> dict: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) UpperCAmelCase__ : Optional[int] = array.dtype UpperCAmelCase__ : Optional[Any] = dtype.byteorder if dtype.byteorder != """=""" else _NATIVE_BYTEORDER UpperCAmelCase__ : List[Any] = dtype.kind UpperCAmelCase__ : Dict = dtype.itemsize UpperCAmelCase__ : Union[str, Any] = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: UpperCAmelCase__ : Union[str, Any] = np.dtype("""|u1""" ) if dtype_kind not in ["u", "i"]: raise TypeError( F"Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays." ) if dtype is not dest_dtype: warnings.warn(F"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'" ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: UpperCAmelCase__ : List[Any] = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: UpperCAmelCase__ : int = dtype_byteorder + dtype_kind + str(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = np.dtype(__lowerCamelCase ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'" ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F"Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}" ) UpperCAmelCase__ : int = PIL.Image.fromarray(array.astype(__lowerCamelCase ) ) return {"path": None, "bytes": image_to_bytes(__lowerCamelCase )} def _lowerCamelCase ( __lowerCamelCase ) -> List[dict]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if objs: UpperCAmelCase__ , UpperCAmelCase__ : List[str] = first_non_null_value(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__lowerCamelCase , np.ndarray ): UpperCAmelCase__ : str = no_op_if_value_is_null(__lowerCamelCase ) return [obj_to_image_dict_func(__lowerCamelCase ) for obj in objs] elif isinstance(__lowerCamelCase , PIL.Image.Image ): UpperCAmelCase__ : List[str] = no_op_if_value_is_null(__lowerCamelCase ) return [obj_to_image_dict_func(__lowerCamelCase ) for obj in objs] else: return objs else: return objs
79
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys a : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
def snake_case ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' if height >= 1: move_tower(height - 1 , lowerCamelCase , lowerCamelCase , lowerCamelCase ) move_disk(lowerCamelCase , lowerCamelCase ) move_tower(height - 1 , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' print("""moving disk from""" , lowerCamelCase , """to""" , lowerCamelCase ) def snake_case ( ): '''simple docstring''' __lowercase = int(input("""Height of hanoi: """ ).strip() ) move_tower(lowerCamelCase , """A""" , """B""" , """C""" ) if __name__ == "__main__": main()
80
def lowerCamelCase__ ( __lowerCamelCase : int ): if num <= 0: raise ValueError("""Input must be a positive integer""" ) __UpperCAmelCase : int = [True] * (num + 1) __UpperCAmelCase : Tuple = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCamelCase ): __UpperCAmelCase : str = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() a : Any = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
63
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _snake_case : str = { "configuration_vision_text_dual_encoder": ["VisionTextDualEncoderConfig"], "processing_vision_text_dual_encoder": ["VisionTextDualEncoderProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = ["VisionTextDualEncoderModel"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ["FlaxVisionTextDualEncoderModel"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["TFVisionTextDualEncoderModel"] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys _snake_case : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
81
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Union[str, Any] = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class a ( lowercase__ ): """simple docstring""" a : Optional[int] = 'git_vision_model' def __init__( self : str , __lowercase : List[str]=768 , __lowercase : List[str]=3072 , __lowercase : List[Any]=12 , __lowercase : Dict=12 , __lowercase : int=3 , __lowercase : Any=224 , __lowercase : Optional[int]=16 , __lowercase : Dict="quick_gelu" , __lowercase : Any=1e-5 , __lowercase : str=0.0 , __lowercase : int=0.02 , **__lowercase : int , ) -> List[str]: super().__init__(**__lowercase ) __UpperCAmelCase : int = hidden_size __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Dict = num_attention_heads __UpperCAmelCase : int = num_channels __UpperCAmelCase : str = patch_size __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : int = initializer_range __UpperCAmelCase : Tuple = attention_dropout __UpperCAmelCase : Optional[int] = layer_norm_eps __UpperCAmelCase : List[Any] = hidden_act @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : Union[str, os.PathLike] , **__lowercase : List[str] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__lowercase ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("""model_type""" ) == "git": __UpperCAmelCase : str = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a ( lowercase__ ): """simple docstring""" a : List[str] = 'git' def __init__( self : Optional[int] , __lowercase : List[Any]=None , __lowercase : Tuple=30522 , __lowercase : str=768 , __lowercase : Optional[int]=6 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[int]=3072 , __lowercase : List[str]="gelu" , __lowercase : Tuple=0.1 , __lowercase : Optional[int]=0.1 , __lowercase : Optional[int]=1024 , __lowercase : Union[str, Any]=0.02 , __lowercase : Optional[Any]=1e-1_2 , __lowercase : List[Any]=0 , __lowercase : Dict="absolute" , __lowercase : Dict=True , __lowercase : Any=False , __lowercase : Optional[int]=101 , __lowercase : str=102 , __lowercase : Union[str, Any]=None , **__lowercase : Dict , ) -> Tuple: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: __UpperCAmelCase : Optional[int] = {} logger.info("""vision_config is None. initializing the GitVisionConfig with default values.""" ) __UpperCAmelCase : Tuple = GitVisionConfig(**__lowercase ) __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Union[str, Any] = hidden_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : List[str] = hidden_act __UpperCAmelCase : str = intermediate_size __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : int = max_position_embeddings __UpperCAmelCase : str = initializer_range __UpperCAmelCase : str = layer_norm_eps __UpperCAmelCase : Union[str, Any] = position_embedding_type __UpperCAmelCase : Dict = use_cache __UpperCAmelCase : int = tie_word_embeddings __UpperCAmelCase : Optional[int] = num_image_with_embedding __UpperCAmelCase : Optional[int] = bos_token_id __UpperCAmelCase : List[Any] = eos_token_id def UpperCAmelCase ( self : str ) -> int: __UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : List[str] = self.vision_config.to_dict() __UpperCAmelCase : Union[str, Any] = self.__class__.model_type return output
63
0
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowercase__ : '''simple docstring''' def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : str ) -> Dict: '''simple docstring''' return None class lowercase__ : '''simple docstring''' def lowercase__ ( self : Tuple , _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Union[str, Any] ) -> Dict: '''simple docstring''' return None class lowercase__ ( unittest.TestCase ): '''simple docstring''' UpperCamelCase = [ # (model_name, model_kwargs) ('''bert-base-cased''', {}), ('''gpt2''', {'''use_cache''': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def lowercase__ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(_UpperCAmelCase , "tf" , 12 , **_UpperCAmelCase ) @require_torch @slow def lowercase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(_UpperCAmelCase , "pt" , 12 , **_UpperCAmelCase ) @require_torch @slow def lowercase__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' from transformers import BertModel UpperCAmelCase_ = ["[UNK]", "[SEP]", "[CLS]", "[PAD]", "[MASK]", "some", "other", "words"] with NamedTemporaryFile(mode="w+t" ) as vocab_file: vocab_file.write("\n".join(_UpperCAmelCase ) ) vocab_file.flush() UpperCAmelCase_ = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: UpperCAmelCase_ = BertModel(BertConfig(vocab_size=len(_UpperCAmelCase ) ) ) model.save_pretrained(_UpperCAmelCase ) self._test_export(_UpperCAmelCase , "pt" , 12 , _UpperCAmelCase ) @require_tf @slow def lowercase__ ( self : str ) -> Any: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: UpperCAmelCase_ = self._test_export(_UpperCAmelCase , "tf" , 12 , **_UpperCAmelCase ) UpperCAmelCase_ = quantize(Path(_UpperCAmelCase ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(_UpperCAmelCase ).stat().st_size: self.fail("Quantized model is bigger than initial ONNX model" ) @require_torch @slow def lowercase__ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: UpperCAmelCase_ = self._test_export(_UpperCAmelCase , "pt" , 12 , **_UpperCAmelCase ) UpperCAmelCase_ = quantize(_UpperCAmelCase ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(_UpperCAmelCase ).stat().st_size: self.fail("Quantized model is bigger than initial ONNX model" ) def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any]=None , **_UpperCAmelCase : Dict ) -> Optional[Any]: '''simple docstring''' try: # Compute path with TemporaryDirectory() as tempdir: UpperCAmelCase_ = Path(_UpperCAmelCase ).joinpath("model.onnx" ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ) return path except Exception as e: self.fail(_UpperCAmelCase ) @require_torch @require_tokenizers @slow def lowercase__ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' from transformers import BertModel UpperCAmelCase_ = BertModel(BertConfig.from_pretrained("lysandre/tiny-bert-random" ) ) UpperCAmelCase_ = BertTokenizerFast.from_pretrained("lysandre/tiny-bert-random" ) self._test_infer_dynamic_axis(_UpperCAmelCase , _UpperCAmelCase , "pt" ) @require_tf @require_tokenizers @slow def lowercase__ ( self : Optional[Any] ) -> List[Any]: '''simple docstring''' from transformers import TFBertModel UpperCAmelCase_ = TFBertModel(BertConfig.from_pretrained("lysandre/tiny-bert-random" ) ) UpperCAmelCase_ = BertTokenizerFast.from_pretrained("lysandre/tiny-bert-random" ) self._test_infer_dynamic_axis(_UpperCAmelCase , _UpperCAmelCase , "tf" ) def lowercase__ ( self : Tuple , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ = FeatureExtractionPipeline(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = ["input_ids", "token_type_ids", "attention_mask", "output_0", "output_1"] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = infer_shapes(_UpperCAmelCase , _UpperCAmelCase ) # Assert all variables are present self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , _UpperCAmelCase ) self.assertSequenceEqual(variable_names[3:] , _UpperCAmelCase ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: "batch", 1: "sequence"} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes["output_0"] , {0: "batch", 1: "sequence"} ) self.assertDictEqual(shapes["output_1"] , {0: "batch"} ) def lowercase__ ( self : Any ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = ["input_ids", "attention_mask", "token_type_ids"] UpperCAmelCase_ = {"input_ids": [1, 2, 3, 4], "attention_mask": [0, 0, 0, 0], "token_type_ids": [1, 1, 1, 1]} UpperCAmelCase_ , UpperCAmelCase_ = ensure_valid_input(FuncContiguousArgs() , _UpperCAmelCase , _UpperCAmelCase ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(_UpperCAmelCase ) , 3 ) # Should have exactly the same input names self.assertEqual(set(_UpperCAmelCase ) , set(_UpperCAmelCase ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(_UpperCAmelCase , (tokens["input_ids"], tokens["token_type_ids"], tokens["attention_mask"]) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) UpperCAmelCase_ , UpperCAmelCase_ = ensure_valid_input(FuncNonContiguousArgs() , _UpperCAmelCase , _UpperCAmelCase ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(_UpperCAmelCase ) , 1 ) self.assertEqual(len(_UpperCAmelCase ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens["input_ids"] ) self.assertEqual(ordered_input_names[0] , "input_ids" ) def lowercase__ ( self : int ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = generate_identified_filename(Path("/home/something/my_fake_model.onnx" ) , "-test" ) self.assertEqual("/home/something/my_fake_model-test.onnx" , generated.as_posix() )
82
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = BarthezTokenizer a : Any = BarthezTokenizerFast a : Union[str, Any] = True a : Union[str, Any] = True def UpperCAmelCase ( self : Dict ) -> Any: super().setUp() __UpperCAmelCase : Optional[int] = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=__lowercase ) __UpperCAmelCase : str = tokenizer def UpperCAmelCase ( self : Optional[int] ) -> Tuple: __UpperCAmelCase : Dict = """<pad>""" __UpperCAmelCase : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> str: __UpperCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__lowercase ) , 101122 ) def UpperCAmelCase ( self : Any ) -> List[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __UpperCAmelCase : str = [0, 57, 3018, 70307, 91, 2] __UpperCAmelCase : List[Any] = self.tokenizer( __lowercase , max_length=len(__lowercase ) , padding=__lowercase , truncation=__lowercase , return_tensors="""pt""" ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __UpperCAmelCase : int = batch.input_ids.tolist()[0] self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> Tuple: if not self.test_rust_tokenizer: return __UpperCAmelCase : Union[str, Any] = self.get_tokenizer() __UpperCAmelCase : Optional[Any] = self.get_rust_tokenizer() __UpperCAmelCase : int = """I was born in 92000, and this is falsé.""" __UpperCAmelCase : Union[str, Any] = tokenizer.tokenize(__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = self.get_rust_tokenizer() __UpperCAmelCase : str = tokenizer.encode(__lowercase ) __UpperCAmelCase : Tuple = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: # fmt: off __UpperCAmelCase : str = {"""input_ids""": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __UpperCAmelCase : int = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=__lowercase , )
63
0
"""simple docstring""" from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = TypeVar('''DatasetType''', Dataset, IterableDataset) def snake_case_ ( A_ : List[DatasetType], A_ : Optional[List[float]] = None, A_ : Optional[int] = None, A_ : Optional[DatasetInfo] = None, A_ : Optional[NamedSplit] = None, A_ : Literal["first_exhausted", "all_exhausted"] = "first_exhausted", ): '''simple docstring''' from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(A_ ): if not isinstance(A_, (Dataset, IterableDataset) ): if isinstance(A_, (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ''' '''is an empty dataset dictionary.''' ) raise ValueError( F'''Dataset at position {i} has at least one split: {list(A_ )}\n''' F'''Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(A_ ) )}\']''' ) raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(A_ ).__name__}.''' ) if i == 0: _lowerCamelCase , _lowerCamelCase : Any = ( (Dataset, IterableDataset) if isinstance(A_, A_ ) else (IterableDataset, Dataset) ) elif not isinstance(A_, A_ ): raise ValueError( F'''Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.''' ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F'''{stopping_strategy} is not supported. Please enter a valid stopping_strategy.''' ) if dataset_type is Dataset: return _interleave_map_style_datasets( A_, A_, A_, info=A_, split=A_, stopping_strategy=A_ ) else: return _interleave_iterable_datasets( A_, A_, A_, info=A_, split=A_, stopping_strategy=A_ ) def snake_case_ ( A_ : List[DatasetType], A_ : Optional[DatasetInfo] = None, A_ : Optional[NamedSplit] = None, A_ : int = 0, ): '''simple docstring''' if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(A_ ): if not isinstance(A_, (Dataset, IterableDataset) ): if isinstance(A_, (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ''' '''is an empty dataset dictionary.''' ) raise ValueError( F'''Dataset at position {i} has at least one split: {list(A_ )}\n''' F'''Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(A_ ) )}\']''' ) raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(A_ ).__name__}.''' ) if i == 0: _lowerCamelCase , _lowerCamelCase : str = ( (Dataset, IterableDataset) if isinstance(A_, A_ ) else (IterableDataset, Dataset) ) elif not isinstance(A_, A_ ): raise ValueError( F'''Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.''' ) if dataset_type is Dataset: return _concatenate_map_style_datasets(A_, info=A_, split=A_, axis=A_ ) else: return _concatenate_iterable_datasets(A_, info=A_, split=A_, axis=A_ )
83
from __future__ import annotations import math def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : bool , __lowerCamelCase : list[int] , __lowerCamelCase : float ): if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(__lowerCamelCase ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = [90, 23, 6, 33, 21, 65, 123, 34423] __UpperCAmelCase : str = math.log(len(__lowerCamelCase ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[str] = ["""image_processor""", """tokenizer"""] _UpperCamelCase : Any = """OwlViTImageProcessor""" _UpperCamelCase : Dict = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , snake_case=None , snake_case=None , **snake_case ): lowercase = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case , ) lowercase = kwargs.pop('feature_extractor' ) lowercase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case , snake_case ) def __call__( self , snake_case=None , snake_case=None , snake_case=None , snake_case="max_length" , snake_case="np" , **snake_case ): if text is None and query_images is None and images is None: raise ValueError( 'You have to specify at least one text or query image or image. All three cannot be none.' ) if text is not None: if isinstance(snake_case , snake_case ) or (isinstance(snake_case , snake_case ) and not isinstance(text[0] , snake_case )): lowercase = [self.tokenizer(snake_case , padding=snake_case , return_tensors=snake_case , **snake_case )] elif isinstance(snake_case , snake_case ) and isinstance(text[0] , snake_case ): lowercase = [] # Maximum number of queries across batch lowercase = max([len(snake_case ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(snake_case ) != max_num_queries: lowercase = t + [' '] * (max_num_queries - len(snake_case )) lowercase = self.tokenizer(snake_case , padding=snake_case , return_tensors=snake_case , **snake_case ) encodings.append(snake_case ) else: raise TypeError('Input text should be a string, a list of strings or a nested list of strings' ) if return_tensors == "np": lowercase = np.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = np.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp lowercase = jnp.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = jnp.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch lowercase = torch.cat([encoding['input_ids'] for encoding in encodings] , dim=0 ) lowercase = torch.cat([encoding['attention_mask'] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf lowercase = tf.stack([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = tf.stack([encoding['attention_mask'] for encoding in encodings] , axis=0 ) else: raise ValueError('Target return tensor type could not be returned' ) lowercase = BatchEncoding() lowercase = input_ids lowercase = attention_mask if query_images is not None: lowercase = BatchEncoding() lowercase = self.image_processor( snake_case , return_tensors=snake_case , **snake_case ).pixel_values lowercase = query_pixel_values if images is not None: lowercase = self.image_processor(snake_case , return_tensors=snake_case , **snake_case ) if text is not None and images is not None: lowercase = image_features.pixel_values return encoding elif query_images is not None and images is not None: lowercase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**snake_case ) , tensor_type=snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process_object_detection(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process_image_guided_detection(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.batch_decode(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.decode(*snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self ): warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self ): warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case , ) return self.image_processor
84
from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class a ( lowercase__ ): """simple docstring""" a : Optional[Any] = 'openai-gpt' a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __lowercase : Tuple=40478 , __lowercase : Tuple=512 , __lowercase : int=768 , __lowercase : Dict=12 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[Any]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Optional[Any]=0.1 , __lowercase : Tuple=0.1 , __lowercase : Dict=1e-5 , __lowercase : Any=0.02 , __lowercase : List[str]="cls_index" , __lowercase : str=True , __lowercase : Dict=None , __lowercase : str=True , __lowercase : List[str]=0.1 , **__lowercase : List[Any] , ) -> List[Any]: __UpperCAmelCase : Optional[int] = vocab_size __UpperCAmelCase : Optional[Any] = n_positions __UpperCAmelCase : Optional[int] = n_embd __UpperCAmelCase : str = n_layer __UpperCAmelCase : Any = n_head __UpperCAmelCase : Tuple = afn __UpperCAmelCase : Any = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : str = attn_pdrop __UpperCAmelCase : str = layer_norm_epsilon __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : Optional[int] = summary_type __UpperCAmelCase : Optional[Any] = summary_use_proj __UpperCAmelCase : List[Any] = summary_activation __UpperCAmelCase : Union[str, Any] = summary_first_dropout __UpperCAmelCase : Dict = summary_proj_to_labels super().__init__(**__lowercase )
63
0
from random import randint from tempfile import TemporaryFile import numpy as np def _a ( lowercase__ : List[Any] , lowercase__ : List[Any] , lowercase__ : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 0 if start < end: SCREAMING_SNAKE_CASE__ : List[str] = randint(lowercase__ , lowercase__ ) SCREAMING_SNAKE_CASE__ : Dict = a[end] SCREAMING_SNAKE_CASE__ : Tuple = a[pivot] SCREAMING_SNAKE_CASE__ : Optional[Any] = temp SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] = _in_place_partition(lowercase__ , lowercase__ , lowercase__ ) count += _in_place_quick_sort(lowercase__ , lowercase__ , p - 1 ) count += _in_place_quick_sort(lowercase__ , p + 1 , lowercase__ ) return count def _a ( lowercase__ : Optional[Any] , lowercase__ : List[str] , lowercase__ : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = 0 SCREAMING_SNAKE_CASE__ : Any = randint(lowercase__ , lowercase__ ) SCREAMING_SNAKE_CASE__ : List[Any] = a[end] SCREAMING_SNAKE_CASE__ : List[Any] = a[pivot] SCREAMING_SNAKE_CASE__ : Any = temp SCREAMING_SNAKE_CASE__ : Optional[int] = start - 1 for index in range(lowercase__ , lowercase__ ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value SCREAMING_SNAKE_CASE__ : Any = new_pivot_index + 1 SCREAMING_SNAKE_CASE__ : List[Any] = a[new_pivot_index] SCREAMING_SNAKE_CASE__ : List[Any] = a[index] SCREAMING_SNAKE_CASE__ : Optional[int] = temp SCREAMING_SNAKE_CASE__ : Optional[int] = a[new_pivot_index + 1] SCREAMING_SNAKE_CASE__ : Union[str, Any] = a[end] SCREAMING_SNAKE_CASE__ : Dict = temp return new_pivot_index + 1, count SCREAMING_SNAKE_CASE__ : Union[str, Any] = TemporaryFile() SCREAMING_SNAKE_CASE__ : Any = 100 # 1000 elements are to be sorted SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = 0, 1 # mean and standard deviation SCREAMING_SNAKE_CASE__ : int = np.random.normal(mu, sigma, p) np.save(outfile, X) print("The array is") print(X) outfile.seek(0) # using the same array SCREAMING_SNAKE_CASE__ : int = np.load(outfile) SCREAMING_SNAKE_CASE__ : List[str] = len(M) - 1 SCREAMING_SNAKE_CASE__ : Any = _in_place_quick_sort(M, 0, r) print( "No of Comparisons for 100 elements selected from a standard normal distribution" "is :" ) print(z)
85
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = KandinskyVaaInpaintPipeline a : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] a : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] a : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] a : List[Any] = False @property def UpperCAmelCase ( self : int ) -> Dict: return 32 @property def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: return 32 @property def UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def UpperCAmelCase ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return 100 @property def UpperCAmelCase ( self : Dict ) -> Any: torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } __UpperCAmelCase : int = UNetaDConditionModel(**__lowercase ) return model @property def UpperCAmelCase ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self : Any ) -> List[Any]: __UpperCAmelCase : List[str] = self.dummy_unet __UpperCAmelCase : List[str] = self.dummy_movq __UpperCAmelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowercase , ) __UpperCAmelCase : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCAmelCase ( self : str , __lowercase : Tuple , __lowercase : List[str]=0 ) -> Optional[Any]: __UpperCAmelCase : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(__lowercase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask __UpperCAmelCase : Union[str, Any] = np.ones((64, 64) , dtype=np.floataa ) __UpperCAmelCase : List[str] = 0 if str(__lowercase ).startswith("""mps""" ): __UpperCAmelCase : List[str] = torch.manual_seed(__lowercase ) else: __UpperCAmelCase : Optional[int] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __UpperCAmelCase : Optional[Any] = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = """cpu""" __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**__lowercase ) __UpperCAmelCase : Tuple = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(__lowercase ) ) __UpperCAmelCase : Tuple = output.images __UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] __UpperCAmelCase : str = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __UpperCAmelCase : Optional[Any] = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def UpperCAmelCase ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) __UpperCAmelCase : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) __UpperCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa ) __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = """a hat""" __UpperCAmelCase : str = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) __UpperCAmelCase : Any = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) __UpperCAmelCase : int = pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() __UpperCAmelCase : Optional[int] = pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
63
0
def __snake_case ( __UpperCamelCase : int | float | str ): """simple docstring""" try: A_ = float(__UpperCamelCase ) except ValueError: raise ValueError("Please enter a valid number" ) A_ = decimal - int(__UpperCamelCase ) if fractional_part == 0: return int(__UpperCamelCase ), 1 else: A_ = len(str(__UpperCamelCase ).split("." )[1] ) A_ = int(decimal * (10**number_of_frac_digits) ) A_ = 10**number_of_frac_digits A_ , A_ = denominator, numerator while True: A_ = dividend % divisor if remainder == 0: break A_ , A_ = divisor, remainder A_ , A_ = numerator / divisor, denominator / divisor return int(__UpperCamelCase ), int(__UpperCamelCase ) if __name__ == "__main__": print(F"{decimal_to_fraction(2) = }") print(F"{decimal_to_fraction(89.0) = }") print(F"{decimal_to_fraction('67') = }") print(F"{decimal_to_fraction('45.0') = }") print(F"{decimal_to_fraction(1.5) = }") print(F"{decimal_to_fraction('6.25') = }") print(F"{decimal_to_fraction('78td') = }")
86
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a : List[Any] = True except ImportError: a : str = False try: from torch.hub import _get_torch_home a : List[Any] = _get_torch_home() except ImportError: a : int = os.path.expanduser( os.getenv("TORCH_HOME", os.path.join(os.getenv("XDG_CACHE_HOME", "~/.cache"), "torch")) ) a : Optional[Any] = os.path.join(torch_cache_home, "transformers") a : Optional[Any] = "https://cdn.huggingface.co" a : List[str] = "https://s3.amazonaws.com/models.huggingface.co/bert" a : Any = "/".join(str(Path(__file__).resolve()).split("/")[:-1]) a : Optional[int] = os.path.join(PATH, "config.yaml") a : Dict = os.path.join(PATH, "attributes.txt") a : Tuple = os.path.join(PATH, "objects.txt") a : Dict = os.getenv("PYTORCH_PRETRAINED_BERT_CACHE", default_cache_path) a : Dict = os.getenv("PYTORCH_TRANSFORMERS_CACHE", PYTORCH_PRETRAINED_BERT_CACHE) a : Optional[int] = os.getenv("TRANSFORMERS_CACHE", PYTORCH_TRANSFORMERS_CACHE) a : Any = "pytorch_model.bin" a : int = "config.yaml" def lowerCamelCase__ ( __lowerCamelCase : str=OBJECTS , __lowerCamelCase : Union[str, Any]=ATTRIBUTES ): __UpperCAmelCase : Union[str, Any] = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(""",""" )[0].lower().strip() ) __UpperCAmelCase : Dict = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(""",""" )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : List[str] = OrderedDict() with open(__lowerCamelCase , """rb""" ) as f: __UpperCAmelCase : int = pkl.load(__lowerCamelCase )["""model"""] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : List[Any] = ckp.pop(__lowerCamelCase ) if isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : Union[str, Any] = torch.tensor(__lowerCamelCase ) else: assert isinstance(__lowerCamelCase , torch.tensor ), type(__lowerCamelCase ) __UpperCAmelCase : List[str] = v return r class a : """simple docstring""" a : Dict = {} def __init__( self : Dict , __lowercase : dict , __lowercase : str = "root" , __lowercase : Any=0 ) -> Dict: __UpperCAmelCase : List[str] = name __UpperCAmelCase : str = level __UpperCAmelCase : int = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : List[str] = copy.deepcopy(__lowercase ) __UpperCAmelCase : Dict = copy.deepcopy(__lowercase ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : Union[str, Any] = Config(__lowercase , name=__lowercase , level=level + 1 ) __UpperCAmelCase : Union[str, Any] = v setattr(self , __lowercase , __lowercase ) __UpperCAmelCase : Any = d def __repr__( self : Optional[Any] ) -> Optional[int]: return str(list((self._pointer.keys()) ) ) def __setattr__( self : List[str] , __lowercase : List[str] , __lowercase : Tuple ) -> int: __UpperCAmelCase : int = val __UpperCAmelCase : List[str] = val __UpperCAmelCase : Union[str, Any] = key.split(""".""" ) __UpperCAmelCase : List[Any] = len(__lowercase ) - 1 __UpperCAmelCase : List[Any] = self._pointer if len(__lowercase ) > 1: for i, l in enumerate(__lowercase ): if hasattr(self , __lowercase ) and isinstance(getattr(self , __lowercase ) , __lowercase ): setattr(getattr(self , __lowercase ) , """.""".join(levels[i:] ) , __lowercase ) if l == last_level: __UpperCAmelCase : Union[str, Any] = val else: __UpperCAmelCase : Union[str, Any] = pointer[l] def UpperCAmelCase ( self : Tuple ) -> Optional[int]: return self._pointer def UpperCAmelCase ( self : str , __lowercase : Optional[int] , __lowercase : Any ) -> Optional[int]: with open(f"""{file_name}""" , """w""" ) as stream: dump(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] ) -> Any: with open(f"""{file_name}""" , """w""" ) as stream: json.dump(__lowercase , __lowercase ) @staticmethod def UpperCAmelCase ( __lowercase : List[Any] ) -> Optional[Any]: with open(__lowercase ) as stream: __UpperCAmelCase : Any = load(__lowercase , Loader=__lowercase ) return data def __str__( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = """ """ if self._name != "root": __UpperCAmelCase : Optional[Any] = f"""{t * (self._level-1)}{self._name}:\n""" else: __UpperCAmelCase : List[Any] = """""" __UpperCAmelCase : Optional[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__lowercase , __lowercase ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__lowercase ).__name__})\n""" __UpperCAmelCase : int = level return r[:-1] @classmethod def UpperCAmelCase ( cls : List[str] , __lowercase : str , **__lowercase : Any ) -> Any: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) return cls(__lowercase ) @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : str , **__lowercase : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase : int = kwargs.pop("""cache_dir""" , __lowercase ) __UpperCAmelCase : int = kwargs.pop("""force_download""" , __lowercase ) __UpperCAmelCase : str = kwargs.pop("""resume_download""" , __lowercase ) __UpperCAmelCase : Dict = kwargs.pop("""proxies""" , __lowercase ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""local_files_only""" , __lowercase ) if os.path.isdir(__lowercase ): __UpperCAmelCase : List[Any] = os.path.join(__lowercase , __lowercase ) elif os.path.isfile(__lowercase ) or is_remote_url(__lowercase ): __UpperCAmelCase : Tuple = pretrained_model_name_or_path else: __UpperCAmelCase : Optional[int] = hf_bucket_url(__lowercase , filename=__lowercase , use_cdn=__lowercase ) try: # Load from URL or cache if already cached __UpperCAmelCase : Optional[int] = cached_path( __lowercase , cache_dir=__lowercase , force_download=__lowercase , proxies=__lowercase , resume_download=__lowercase , local_files_only=__lowercase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : Optional[int] = Config.load_yaml(__lowercase ) except EnvironmentError: __UpperCAmelCase : str = """Can't load config for""" raise EnvironmentError(__lowercase ) if resolved_config_file == config_file: print("""loading configuration file from path""" ) else: print("""loading configuration file cache""" ) return Config.load_yaml(__lowercase ), kwargs def lowerCamelCase__ ( __lowerCamelCase : Dict ): __UpperCAmelCase : Optional[int] = torch.load("""dump.pt""" , map_location=in_tensor.device ) __UpperCAmelCase : Tuple = in_tensor.numpy() __UpperCAmelCase : Optional[int] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception("""tensors are all good""" ) # Hugging face functions below def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Tuple = urlparse(__lowerCamelCase ) return parsed.scheme in ("http", "https") def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : int=True ): __UpperCAmelCase : int = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : Optional[int] = """/""" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=None , ): __UpperCAmelCase : Optional[int] = """python/{}""".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + "; ".join("""{}/{}""".format(__lowerCamelCase , __lowerCamelCase ) for k, v in user_agent.items() ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + user_agent __UpperCAmelCase : List[str] = {"""user-agent""": ua} if resume_size > 0: __UpperCAmelCase : Union[str, Any] = """bytes=%d-""" % (resume_size,) __UpperCAmelCase : Union[str, Any] = requests.get(__lowerCamelCase , stream=__lowerCamelCase , proxies=__lowerCamelCase , headers=__lowerCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : List[str] = response.headers.get("""Content-Length""" ) __UpperCAmelCase : str = resume_size + int(__lowerCamelCase ) if content_length is not None else None __UpperCAmelCase : List[Any] = tqdm( unit="""B""" , unit_scale=__lowerCamelCase , total=__lowerCamelCase , initial=__lowerCamelCase , desc="""Downloading""" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(__lowerCamelCase ) ) temp_file.write(__lowerCamelCase ) progress.close() def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=10 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=None , __lowerCamelCase : List[str]=False , ): if cache_dir is None: __UpperCAmelCase : Optional[Any] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[str] = str(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) __UpperCAmelCase : List[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(__lowerCamelCase , allow_redirects=__lowerCamelCase , proxies=__lowerCamelCase , timeout=__lowerCamelCase ) if response.status_code == 200: __UpperCAmelCase : Dict = response.headers.get("""ETag""" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : List[str] = url_to_filename(__lowerCamelCase , __lowerCamelCase ) # get cache path to put the file __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , __lowerCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(__lowerCamelCase ): return cache_path else: __UpperCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(__lowerCamelCase ) , filename + """.*""" ) if not file.endswith(""".json""" ) and not file.endswith(""".lock""" ) ] if len(__lowerCamelCase ) > 0: return os.path.join(__lowerCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( """Cannot find the requested files in the cached path and outgoing traffic has been""" """ disabled. To enable model look-ups and downloads online, set 'local_files_only'""" """ to False.""" ) return None # From now on, etag is not None. if os.path.exists(__lowerCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : str = cache_path + """.lock""" with FileLock(__lowerCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(__lowerCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : int = cache_path + """.incomplete""" @contextmanager def _resumable_file_manager(): with open(__lowerCamelCase , """a+b""" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(__lowerCamelCase ): __UpperCAmelCase : List[Any] = os.stat(__lowerCamelCase ).st_size else: __UpperCAmelCase : List[Any] = 0 else: __UpperCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=__lowerCamelCase , delete=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( """%s not found in cache or force_download set to True, downloading to %s""" , __lowerCamelCase , temp_file.name , ) http_get( __lowerCamelCase , __lowerCamelCase , proxies=__lowerCamelCase , resume_size=__lowerCamelCase , user_agent=__lowerCamelCase , ) os.replace(temp_file.name , __lowerCamelCase ) __UpperCAmelCase : Any = {"""url""": url, """etag""": etag} __UpperCAmelCase : Union[str, Any] = cache_path + """.json""" with open(__lowerCamelCase , """w""" ) as meta_file: json.dump(__lowerCamelCase , __lowerCamelCase ) return cache_path def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any]=None ): __UpperCAmelCase : Tuple = url.encode("""utf-8""" ) __UpperCAmelCase : Optional[Any] = shaaaa(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = url_hash.hexdigest() if etag: __UpperCAmelCase : int = etag.encode("""utf-8""" ) __UpperCAmelCase : List[str] = shaaaa(__lowerCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(""".h5""" ): filename += ".h5" return filename def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : int=None , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=False , ): if cache_dir is None: __UpperCAmelCase : List[str] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Any = str(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) if is_remote_url(__lowerCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : Tuple = get_from_cache( __lowerCamelCase , cache_dir=__lowerCamelCase , force_download=__lowerCamelCase , proxies=__lowerCamelCase , resume_download=__lowerCamelCase , user_agent=__lowerCamelCase , local_files_only=__lowerCamelCase , ) elif os.path.exists(__lowerCamelCase ): # File, and it exists. __UpperCAmelCase : Tuple = url_or_filename elif urlparse(__lowerCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("""file {} not found""".format(__lowerCamelCase ) ) else: # Something unknown raise ValueError("""unable to parse {} as a URL or as a local path""".format(__lowerCamelCase ) ) if extract_compressed_file: if not is_zipfile(__lowerCamelCase ) and not tarfile.is_tarfile(__lowerCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : int = os.path.split(__lowerCamelCase ) __UpperCAmelCase : Any = output_file.replace(""".""" , """-""" ) + """-extracted""" __UpperCAmelCase : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isdir(__lowerCamelCase ) and os.listdir(__lowerCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : str = output_path + """.lock""" with FileLock(__lowerCamelCase ): shutil.rmtree(__lowerCamelCase , ignore_errors=__lowerCamelCase ) os.makedirs(__lowerCamelCase ) if is_zipfile(__lowerCamelCase ): with ZipFile(__lowerCamelCase , """r""" ) as zip_file: zip_file.extractall(__lowerCamelCase ) zip_file.close() elif tarfile.is_tarfile(__lowerCamelCase ): __UpperCAmelCase : Any = tarfile.open(__lowerCamelCase ) tar_file.extractall(__lowerCamelCase ) tar_file.close() else: raise EnvironmentError("""Archive format of {} could not be identified""".format(__lowerCamelCase ) ) return output_path_extracted return output_path def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : int="," ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): with open(__lowerCamelCase ) as f: __UpperCAmelCase : List[Any] = eval(f.read() ) else: __UpperCAmelCase : List[str] = requests.get(__lowerCamelCase ) try: __UpperCAmelCase : int = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : str = eval(__lowerCamelCase ) except Exception: __UpperCAmelCase : List[Any] = data.split("""\n""" ) req.close() return data def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : Optional[int] = requests.get(__lowerCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCamelCase__ ( __lowerCamelCase : str ): __UpperCAmelCase : int = url.split("""/""" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(__lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as stream: __UpperCAmelCase : List[str] = pkl.load(__lowerCamelCase ) __UpperCAmelCase : Dict = weights.pop("""model""" ) __UpperCAmelCase : Union[str, Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(__lowerCamelCase ) if "running_var" in k: __UpperCAmelCase : Optional[int] = torch.tensor([0] ) __UpperCAmelCase : Tuple = k.replace("""running_var""" , """num_batches_tracked""" ) __UpperCAmelCase : Any = zero return new def lowerCamelCase__ ( ): print(f"""{os.path.abspath(os.path.join(__lowerCamelCase , os.pardir ) )}/demo.ipynb""" ) def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]="RGB" ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): __UpperCAmelCase : List[str] = cva.imread(__lowerCamelCase ) else: __UpperCAmelCase : int = get_image_from_url(__lowerCamelCase ) assert img is not None, f"""could not connect to: {im}""" __UpperCAmelCase : Any = cva.cvtColor(__lowerCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Optional[int] = img[:, :, ::-1] return img def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int=1 ): return (images[i : i + batch] for i in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ))
63
0
from __future__ import annotations _lowerCamelCase : Tuple = tuple[int, int, int] _lowerCamelCase : Union[str, Any] = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase _lowerCamelCase : List[Any] = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" # -------------------------- default selection -------------------------- # rotors -------------------------- _lowerCamelCase : str = """EGZWVONAHDCLFQMSIPJBYUKXTR""" _lowerCamelCase : str = """FOBHMDKEXQNRAULPGSJVTYICZW""" _lowerCamelCase : List[Any] = """ZJXESIUQLHAVRMDOYGTNFWPBKC""" # reflector -------------------------- _lowerCamelCase : Union[str, Any] = { """A""": """N""", """N""": """A""", """B""": """O""", """O""": """B""", """C""": """P""", """P""": """C""", """D""": """Q""", """Q""": """D""", """E""": """R""", """R""": """E""", """F""": """S""", """S""": """F""", """G""": """T""", """T""": """G""", """H""": """U""", """U""": """H""", """I""": """V""", """V""": """I""", """J""": """W""", """W""": """J""", """K""": """X""", """X""": """K""", """L""": """Y""", """Y""": """L""", """M""": """Z""", """Z""": """M""", } # -------------------------- extra rotors -------------------------- _lowerCamelCase : Tuple = """RMDJXFUWGISLHVTCQNKYPBEZOA""" _lowerCamelCase : Any = """SGLCPQWZHKXAREONTFBVIYJUDM""" _lowerCamelCase : Optional[Any] = """HVSICLTYKQUBXDWAJZOMFGPREN""" _lowerCamelCase : Dict = """RZWQHFMVDBKICJLNTUXAGYPSOE""" _lowerCamelCase : List[str] = """LFKIJODBEGAMQPXVUHYSTCZRWN""" _lowerCamelCase : Optional[int] = """KOAEGVDHXPQZMLFTYWJNBRCIUS""" def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: """simple docstring""" if (unique_rotsel := len(set(lowercase_ ) )) < 3: A__ = f"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(lowercase_ ) # Checks if rotor positions are valid A__ , A__ , A__ = rotpos if not 0 < rotorposa <= len(lowercase_ ): A__ = f"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(lowercase_ ) if not 0 < rotorposa <= len(lowercase_ ): A__ = f"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(lowercase_ ) if not 0 < rotorposa <= len(lowercase_ ): A__ = f"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(lowercase_ ) # Validates string and returns dict A__ = _plugboard(lowercase_ ) return rotpos, rotsel, pbdict def SCREAMING_SNAKE_CASE ( lowercase_ ) -> dict[str, str]: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ): A__ = f"""Plugboard setting isn't type string ({type(lowercase_ )})""" raise TypeError(lowercase_ ) elif len(lowercase_ ) % 2 != 0: A__ = f"""Odd number of symbols ({len(lowercase_ )})""" raise Exception(lowercase_ ) elif pbstring == "": return {} pbstring.replace(''' ''' , '''''' ) # Checks if all characters are unique A__ = set() for i in pbstring: if i not in abc: A__ = f"""'{i}' not in list of symbols""" raise Exception(lowercase_ ) elif i in tmppbl: A__ = f"""Duplicate symbol ({i})""" raise Exception(lowercase_ ) else: tmppbl.add(lowercase_ ) del tmppbl # Created the dictionary A__ = {} for j in range(0 , len(lowercase_ ) - 1 , 2 ): A__ = pbstring[j + 1] A__ = pbstring[j] return pb def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ = (rotora, rotora, rotora) , lowercase_ = "" , ) -> str: """simple docstring""" A__ = text.upper() A__ , A__ , A__ = _validator( lowercase_ , lowercase_ , plugb.upper() ) A__ , A__ , A__ = rotor_position A__ , A__ , A__ = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 A__ = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: A__ = plugboard[symbol] # rotor ra -------------------------- A__ = abc.index(lowercase_ ) + rotorposa A__ = rotora[index % len(lowercase_ )] # rotor rb -------------------------- A__ = abc.index(lowercase_ ) + rotorposa A__ = rotora[index % len(lowercase_ )] # rotor rc -------------------------- A__ = abc.index(lowercase_ ) + rotorposa A__ = rotora[index % len(lowercase_ )] # reflector -------------------------- # this is the reason you don't need another machine to decipher A__ = reflector[symbol] # 2nd rotors A__ = abc[rotora.index(lowercase_ ) - rotorposa] A__ = abc[rotora.index(lowercase_ ) - rotorposa] A__ = abc[rotora.index(lowercase_ ) - rotorposa] # 2nd plugboard if symbol in plugboard: A__ = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(lowercase_ ): A__ = 0 rotorposa += 1 if rotorposa >= len(lowercase_ ): A__ = 0 rotorposa += 1 if rotorposa >= len(lowercase_ ): A__ = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(lowercase_ ) return "".join(lowercase_ ) if __name__ == "__main__": _lowerCamelCase : Union[str, Any] = """This is my Python script that emulates the Enigma machine from WWII.""" _lowerCamelCase : str = (1, 1, 1) _lowerCamelCase : Optional[Any] = """pictures""" _lowerCamelCase : List[str] = (rotora, rotora, rotora) _lowerCamelCase : Any = enigma(message, rotor_pos, rotor_sel, pb) print("""Encrypted message:""", en) print("""Decrypted message:""", enigma(en, rotor_pos, rotor_sel, pb))
87
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Any=13 , __lowercase : Optional[int]=7 , __lowercase : str=True , __lowercase : Optional[Any]=True , __lowercase : int=True , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : int=5 , __lowercase : Tuple=4 , __lowercase : str=37 , __lowercase : Optional[int]="gelu" , __lowercase : Tuple=0.1 , __lowercase : str=0.1 , __lowercase : Dict=512 , __lowercase : List[Any]=16 , __lowercase : Dict=2 , __lowercase : Union[str, Any]=0.02 , __lowercase : Dict=4 , ) -> int: __UpperCAmelCase : Dict = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : str = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Dict = use_token_type_ids __UpperCAmelCase : Optional[int] = use_labels __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : str = type_sequence_label_size __UpperCAmelCase : List[str] = initializer_range __UpperCAmelCase : Tuple = num_choices def UpperCAmelCase ( self : Dict ) -> Tuple: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : List[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase ( self : Tuple ) -> List[Any]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase ( self : Any ) -> List[str]: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[int] = config_and_inputs __UpperCAmelCase : int = True __UpperCAmelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = True a : List[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : List[str] = FlaxRobertaModelTester(self ) @slow def UpperCAmelCase ( self : str ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase : Union[str, Any] = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowercase ) __UpperCAmelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowercase )
63
0
"""simple docstring""" from ... import PretrainedConfig UpperCAmelCase = { """sijunhe/nezha-cn-base""": """https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json""", } class lowercase__ ( A_ ): __UpperCAmelCase = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP __UpperCAmelCase = '''nezha''' def __init__( self , SCREAMING_SNAKE_CASE=2_1128 , SCREAMING_SNAKE_CASE=768 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=3072 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=512 , SCREAMING_SNAKE_CASE=64 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=0.02 , SCREAMING_SNAKE_CASE=1e-1_2 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=True , **SCREAMING_SNAKE_CASE , ) -> Union[str, Any]: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE , bos_token_id=SCREAMING_SNAKE_CASE , eos_token_id=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) _lowerCamelCase : Dict = vocab_size _lowerCamelCase : Tuple = hidden_size _lowerCamelCase : Dict = num_hidden_layers _lowerCamelCase : List[Any] = num_attention_heads _lowerCamelCase : Any = hidden_act _lowerCamelCase : Dict = intermediate_size _lowerCamelCase : List[str] = hidden_dropout_prob _lowerCamelCase : int = attention_probs_dropout_prob _lowerCamelCase : List[str] = max_position_embeddings _lowerCamelCase : Dict = max_relative_position _lowerCamelCase : Union[str, Any] = type_vocab_size _lowerCamelCase : int = initializer_range _lowerCamelCase : Tuple = layer_norm_eps _lowerCamelCase : Union[str, Any] = classifier_dropout _lowerCamelCase : str = use_cache
88
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging a : Optional[int] = logging.get_logger(__name__) class a ( lowercase__ ): """simple docstring""" a : Tuple = 'linear' a : int = 'cosine' a : Optional[Any] = 'cosine_with_restarts' a : Dict = 'polynomial' a : Tuple = 'constant' a : Dict = 'constant_with_warmup' a : Any = 'piecewise_constant' def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int = -1 ): return LambdaLR(__lowerCamelCase , lambda __lowerCamelCase : 1 , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : str , __lowerCamelCase : int = -1 ): __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Tuple = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __UpperCAmelCase , __UpperCAmelCase : List[str] = rule_str.split(""":""" ) __UpperCAmelCase : Any = int(__lowerCamelCase ) __UpperCAmelCase : List[str] = float(__lowerCamelCase ) __UpperCAmelCase : int = value __UpperCAmelCase : Any = float(rule_list[-1] ) def create_rules_function(__lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): def rule_func(__lowerCamelCase : int ) -> float: __UpperCAmelCase : Tuple = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=-1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float = 0.5 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Dict ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Tuple = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int = 1 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Union[str, Any] ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Union[str, Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=1E-7 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : int=-1 ): __UpperCAmelCase : Tuple = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(f"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __UpperCAmelCase : Optional[Any] = lr_init - lr_end __UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps __UpperCAmelCase : int = 1 - (current_step - num_warmup_steps) / decay_steps __UpperCAmelCase : Optional[int] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a : int = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCamelCase__ ( __lowerCamelCase : Union[str, SchedulerType] , __lowerCamelCase : Optimizer , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 1 , __lowerCamelCase : float = 1.0 , __lowerCamelCase : int = -1 , ): __UpperCAmelCase : Union[str, Any] = SchedulerType(__lowerCamelCase ) __UpperCAmelCase : int = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
63
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
89
from math import pi, sqrt def lowerCamelCase__ ( __lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCamelCase__ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() a : Optional[int] = 1.0 while num: a : List[str] = float(input("Gamma of: ")) print(f"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
63
0
'''simple docstring''' from __future__ import annotations import queue class a__ : '''simple docstring''' def __init__( self , lowerCamelCase_ ) -> int: lowerCAmelCase__ = data lowerCAmelCase__ = None lowerCAmelCase__ = None def _snake_case ( ) -> TreeNode: print('''\n********Press N to stop entering at any point of time********\n''' ) lowerCAmelCase__ = input('''Enter the value of the root node: ''' ).strip().lower() lowerCAmelCase__ = queue.Queue() lowerCAmelCase__ = TreeNode(int(A ) ) q.put(A ) while not q.empty(): lowerCAmelCase__ = q.get() lowerCAmelCase__ = F"""Enter the left node of {node_found.data}: """ lowerCAmelCase__ = input(A ).strip().lower() or '''n''' if check == "n": return tree_node lowerCAmelCase__ = TreeNode(int(A ) ) lowerCAmelCase__ = left_node q.put(A ) lowerCAmelCase__ = F"""Enter the right node of {node_found.data}: """ lowerCAmelCase__ = input(A ).strip().lower() or '''n''' if check == "n": return tree_node lowerCAmelCase__ = TreeNode(int(A ) ) lowerCAmelCase__ = right_node q.put(A ) raise def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = queue.Queue() q.put(A ) while not q.empty(): lowerCAmelCase__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = queue.Queue() q.put(A ) while not q.empty(): lowerCAmelCase__ = [] while not q.empty(): lowerCAmelCase__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(A ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = [] lowerCAmelCase__ = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(A ) lowerCAmelCase__ = n.left # end of while means current node doesn't have left child lowerCAmelCase__ = stack.pop() # start to traverse its right child lowerCAmelCase__ = n.right def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = [] lowerCAmelCase__ = node while n or stack: while n: stack.append(A ) lowerCAmelCase__ = n.left lowerCAmelCase__ = stack.pop() print(n.data , end=''',''' ) lowerCAmelCase__ = n.right def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ , lowerCAmelCase__ = [], [] lowerCAmelCase__ = node stacka.append(A ) while stacka: # to find the reversed order of post order, store it in stack2 lowerCAmelCase__ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(A ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def _snake_case ( A = "" , A=50 , A="*" ) -> str: if not s: return "\n" + width * char lowerCAmelCase__ , lowerCAmelCase__ = divmod(width - len(A ) - 2 , 2 ) return F"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt('''Binary Tree Traversals''')) __UpperCAmelCase = build_tree() print(prompt('''Pre Order Traversal''')) pre_order(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal''')) in_order(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal''')) post_order(node) print(prompt() + '''\n''') print(prompt('''Level Order Traversal''')) level_order(node) print(prompt() + '''\n''') print(prompt('''Actual Level Order Traversal''')) level_order_actual(node) print('''*''' * 50 + '''\n''') print(prompt('''Pre Order Traversal - Iteration Version''')) pre_order_iter(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal - Iteration Version''')) in_order_iter(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal - Iteration Version''')) post_order_iter(node) print(prompt())
90
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class a : """simple docstring""" a : int a : Node | None = None a : Node | None = None def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = Node(1 ) __UpperCAmelCase : int = Node(2 ) __UpperCAmelCase : Optional[Any] = Node(3 ) __UpperCAmelCase : Dict = Node(4 ) __UpperCAmelCase : Tuple = Node(5 ) return tree def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCamelCase__ ( __lowerCamelCase : Node | None ): __UpperCAmelCase : list[Any] = [] if root is None: return output __UpperCAmelCase : Tuple = deque([root] ) while process_queue: __UpperCAmelCase : Optional[Any] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None ): if root is None: return [] __UpperCAmelCase : list[Sequence[Node | None]] = [] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : int = height(__lowerCamelCase ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : Optional[int] = 0 return output def lowerCamelCase__ ( ): # Main function for testing. __UpperCAmelCase : List[Any] = make_tree() print(f"""In-order Traversal: {inorder(__lowerCamelCase )}""" ) print(f"""Pre-order Traversal: {preorder(__lowerCamelCase )}""" ) print(f"""Post-order Traversal: {postorder(__lowerCamelCase )}""" , """\n""" ) print(f"""Height of Tree: {height(__lowerCamelCase )}""" , """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(__lowerCamelCase ) , """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1 , height(__lowerCamelCase ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(__lowerCamelCase , level=__lowerCamelCase ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor _lowercase = logging.get_logger(__name__) class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' def __init__( self : Any ,*A_ : str ,**A_ : List[str] ) -> None: warnings.warn( 'The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use DeformableDetrImageProcessor instead.' ,A_ ,) super().__init__(*A_ ,**A_ )
91
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[int] = GPTSanJapaneseTokenizer a : Optional[Any] = False a : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off __UpperCAmelCase : Tuple = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __UpperCAmelCase : Dict = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __UpperCAmelCase : Dict = {"""unk_token""": """<unk>"""} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(__lowercase ) ) def UpperCAmelCase ( self : Tuple , **__lowercase : int ) -> Any: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def UpperCAmelCase ( self : str , __lowercase : Union[str, Any] ) -> Any: __UpperCAmelCase : Any = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __UpperCAmelCase : int = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase ( self : List[Any] , __lowercase : Optional[int] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase : int = self.get_input_output_texts(__lowercase ) __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : Dict = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) return text, ids def UpperCAmelCase ( self : int ) -> Optional[Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Dict ) -> Tuple: pass # TODO add if relevant def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : List[str] = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。 こんばんは、㔺界。""" __UpperCAmelCase : Dict = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids without special tokens __UpperCAmelCase : List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids with special tokens __UpperCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __UpperCAmelCase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : int = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : Tuple = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __UpperCAmelCase : int = """こんにちは、、、、世界。こんばんは、、、、世界。""" __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase ) __UpperCAmelCase : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : int ) -> Optional[int]: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : List[Any] = """こんにちは、世界。""" __UpperCAmelCase : Optional[int] = """こんばんは、㔺界。😀""" __UpperCAmelCase : List[Any] = """こんにちは、世界。こんばんは、世界。😀""" __UpperCAmelCase : List[str] = tokenizer.encode(prefix_text + input_text ) __UpperCAmelCase : List[Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __UpperCAmelCase : Any = tokenizer.encode(__lowercase , prefix_text=__lowercase ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowercase ) __UpperCAmelCase : Any = tokenizer.decode(__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。""" __UpperCAmelCase : List[Any] = """こんばんは、㔺界。😀""" __UpperCAmelCase : Union[str, Any] = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : int = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : List[Any] = [1] + [0] * (len_prefix + len_text + 1) __UpperCAmelCase : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] __UpperCAmelCase : List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCAmelCase : Union[str, Any] = tokenizer(prefix_text + input_text ).token_type_ids __UpperCAmelCase : Optional[Any] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __UpperCAmelCase : Tuple = tokenizer(__lowercase , prefix_text=__lowercase ).token_type_ids self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : List[str] ) -> int: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""あンいワ""" ) __UpperCAmelCase : Tuple = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertNotEqual(__lowercase , __lowercase ) self.assertNotEqual(__lowercase , __lowercase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self : List[Any] ) -> List[str]: __UpperCAmelCase : Any = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : List[Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __UpperCAmelCase : int = tokenizer(__lowercase , padding=__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.batch_encode_plus(__lowercase , padding=__lowercase ) # fmt: off __UpperCAmelCase : Optional[int] = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCAmelCase : Tuple = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCAmelCase : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __lowercase ) self.assertListEqual(x_token.token_type_ids , __lowercase ) self.assertListEqual(x_token.attention_mask , __lowercase ) self.assertListEqual(x_token_a.input_ids , __lowercase ) self.assertListEqual(x_token_a.token_type_ids , __lowercase ) self.assertListEqual(x_token_a.attention_mask , __lowercase ) def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase ( self : Any ) -> int: # tokenizer has no padding token pass
63
0
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType UpperCamelCase_ = None UpperCamelCase_ = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image UpperCamelCase_ = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class __SCREAMING_SNAKE_CASE : lowerCamelCase_ = True lowerCamelCase_ = None # Automatically constructed lowerCamelCase_ = "PIL.Image.Image" lowerCamelCase_ = pa.struct({'bytes': pa.binary(), 'path': pa.string()} ) lowerCamelCase_ = field(default='Image' , init=lowercase__ , repr=lowercase__ ) def __call__( self : str ): '''simple docstring''' return self.pa_type def lowerCamelCase_ ( self : List[str] , UpperCAmelCase__ : Union[str, bytes, dict, np.ndarray, "PIL.Image.Image"] ): '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): lowercase : Tuple =np.array(UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): return {"path": value, "bytes": None} elif isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): return {"path": None, "bytes": value} elif isinstance(UpperCAmelCase__ , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(UpperCAmelCase__ ) elif isinstance(UpperCAmelCase__ , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(UpperCAmelCase__ ) elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def lowerCamelCase_ ( self : List[str] , UpperCAmelCase__ : dict , UpperCAmelCase__ : Optional[int]=None ): '''simple docstring''' if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Image(decode=True) instead.''' ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support decoding images, please install \'Pillow\'.''' ) if token_per_repo_id is None: lowercase : Tuple ={} lowercase , lowercase : List[Any] =value['''path'''], value['''bytes'''] if bytes_ is None: if path is None: raise ValueError(F'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(UpperCAmelCase__ ): lowercase : int =PIL.Image.open(UpperCAmelCase__ ) else: lowercase : Optional[Any] =path.split('''::''' )[-1] try: lowercase : Any =string_to_dict(UpperCAmelCase__ , config.HUB_DATASETS_URL )['''repo_id'''] lowercase : Union[str, Any] =token_per_repo_id.get(UpperCAmelCase__ ) except ValueError: lowercase : Any =None with xopen(UpperCAmelCase__ , '''rb''' , use_auth_token=UpperCAmelCase__ ) as f: lowercase : int =BytesIO(f.read() ) lowercase : int =PIL.Image.open(bytes_ ) else: lowercase : List[Any] =PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def lowerCamelCase_ ( self : Optional[Any] ): '''simple docstring''' from .features import Value return ( self if self.decode else { "bytes": Value('''binary''' ), "path": Value('''string''' ), } ) def lowerCamelCase_ ( self : Tuple , UpperCAmelCase__ : Union[pa.StringArray, pa.StructArray, pa.ListArray] ): '''simple docstring''' if pa.types.is_string(storage.type ): lowercase : str =pa.array([None] * len(UpperCAmelCase__ ) , type=pa.binary() ) lowercase : str =pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): lowercase : int =pa.array([None] * len(UpperCAmelCase__ ) , type=pa.string() ) lowercase : Union[str, Any] =pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: lowercase : Union[str, Any] =storage.field('''bytes''' ) else: lowercase : Dict =pa.array([None] * len(UpperCAmelCase__ ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: lowercase : Union[str, Any] =storage.field('''path''' ) else: lowercase : List[str] =pa.array([None] * len(UpperCAmelCase__ ) , type=pa.string() ) lowercase : Optional[Any] =pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): lowercase : Optional[Any] =pa.array( [encode_np_array(np.array(UpperCAmelCase__ ) )['''bytes'''] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) lowercase : int =pa.array([None] * len(UpperCAmelCase__ ) , type=pa.string() ) lowercase : List[str] =pa.StructArray.from_arrays( [bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(UpperCAmelCase__ , self.pa_type ) def lowerCamelCase_ ( self : Optional[int] , UpperCAmelCase__ : pa.StructArray ): '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(UpperCAmelCase__ : Optional[int] ): with xopen(UpperCAmelCase__ , '''rb''' ) as f: lowercase : Union[str, Any] =f.read() return bytes_ lowercase : Optional[Any] =pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) lowercase : List[Any] =pa.array( [os.path.basename(UpperCAmelCase__ ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) lowercase : str =pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(UpperCAmelCase__ , self.pa_type ) def _lowerCAmelCase ( ) -> List[str]: if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() lowercase : List[Any] =list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _lowerCAmelCase ( __magic_name__ : "PIL.Image.Image" ) -> bytes: lowercase : int =BytesIO() if image.format in list_image_compression_formats(): lowercase : Any =image.format else: lowercase : Optional[Any] ='''PNG''' if image.mode in ['''1''', '''L''', '''LA''', '''RGB''', '''RGBA'''] else '''TIFF''' image.save(__magic_name__ , format=__magic_name__ ) return buffer.getvalue() def _lowerCAmelCase ( __magic_name__ : "PIL.Image.Image" ) -> dict: if hasattr(__magic_name__ , '''filename''' ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__magic_name__ )} def _lowerCAmelCase ( __magic_name__ : np.ndarray ) -> dict: if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) lowercase : Optional[int] =array.dtype lowercase : Dict =dtype.byteorder if dtype.byteorder != '''=''' else _NATIVE_BYTEORDER lowercase : Dict =dtype.kind lowercase : int =dtype.itemsize lowercase : int =None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: lowercase : Any =np.dtype('''|u1''' ) if dtype_kind not in ["u", "i"]: raise TypeError( f'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''' ) if dtype is not dest_dtype: warnings.warn(f'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: lowercase : str =dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: lowercase : Optional[Any] =dtype_byteorder + dtype_kind + str(__magic_name__ ) lowercase : Union[str, Any] =np.dtype(__magic_name__ ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(f'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( f'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''' ) lowercase : Dict =PIL.Image.fromarray(array.astype(__magic_name__ ) ) return {"path": None, "bytes": image_to_bytes(__magic_name__ )} def _lowerCAmelCase ( __magic_name__ : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ) -> List[dict]: if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) if objs: lowercase , lowercase : Tuple =first_non_null_value(__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__magic_name__ , np.ndarray ): lowercase : Optional[int] =no_op_if_value_is_null(__magic_name__ ) return [obj_to_image_dict_func(__magic_name__ ) for obj in objs] elif isinstance(__magic_name__ , PIL.Image.Image ): lowercase : Optional[int] =no_op_if_value_is_null(__magic_name__ ) return [obj_to_image_dict_func(__magic_name__ ) for obj in objs] else: return objs else: return objs
92
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a : Dict = logging.get_logger(__name__) @dataclass class a ( lowercase__ ): """simple docstring""" a : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : List[Any] , **__lowercase : Dict ) -> Tuple: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase : List[Any] = deprecated_arg[3:] setattr(self , __lowercase , not kwargs.pop(__lowercase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __UpperCAmelCase : str = kwargs.pop("""torchscript""" , self.torchscript ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) __UpperCAmelCase : Optional[Any] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**__lowercase ) a : bool = field(default=lowercase__ , metadata={'help': 'Trace the models using torchscript'} ) a : bool = field(default=lowercase__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) a : str = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def UpperCAmelCase ( self : Any ) -> Tuple["torch.device", int]: requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: __UpperCAmelCase : str = torch.device("""cpu""" ) __UpperCAmelCase : int = 0 elif is_torch_tpu_available(): __UpperCAmelCase : Tuple = xm.xla_device() __UpperCAmelCase : int = 0 else: __UpperCAmelCase : Dict = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) __UpperCAmelCase : Optional[int] = torch.cuda.device_count() return device, n_gpu @property def UpperCAmelCase ( self : Optional[Any] ) -> str: return is_torch_tpu_available() and self.tpu @property def UpperCAmelCase ( self : List[str] ) -> int: requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCAmelCase ( self : int ) -> "torch.device": requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def UpperCAmelCase ( self : int ) -> List[Any]: requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return self.n_gpu > 0
63
0
"""simple docstring""" import math def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0 , _SCREAMING_SNAKE_CASE = 0 ) ->list: """simple docstring""" lowerCAmelCase__ :List[str] = end or len(_SCREAMING_SNAKE_CASE ) for i in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ :List[Any] = i lowerCAmelCase__ :Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowerCAmelCase__ :str = array[temp_index - 1] temp_index -= 1 lowerCAmelCase__ :Tuple = temp_index_value return array def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->None: # Max Heap """simple docstring""" lowerCAmelCase__ :Any = index lowerCAmelCase__ :Optional[int] = 2 * index + 1 # Left Node lowerCAmelCase__ :int = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowerCAmelCase__ :Dict = left_index if right_index < heap_size and array[largest] < array[right_index]: lowerCAmelCase__ :str = right_index if largest != index: lowerCAmelCase__ , lowerCAmelCase__ :Dict = array[largest], array[index] heapify(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE ) ->list: """simple docstring""" lowerCAmelCase__ :Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) for i in range(n // 2 , -1 , -1 ): heapify(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for i in range(n - 1 , 0 , -1 ): lowerCAmelCase__ , lowerCAmelCase__ :int = array[0], array[i] heapify(_SCREAMING_SNAKE_CASE , 0 , _SCREAMING_SNAKE_CASE ) return array def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" lowerCAmelCase__ :Optional[int] = low lowerCAmelCase__ :int = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowerCAmelCase__ , lowerCAmelCase__ :List[str] = array[j], array[i] i += 1 def __A (_SCREAMING_SNAKE_CASE ) ->list: """simple docstring""" if len(_SCREAMING_SNAKE_CASE ) == 0: return array lowerCAmelCase__ :Dict = 2 * math.ceil(math.loga(len(_SCREAMING_SNAKE_CASE ) ) ) lowerCAmelCase__ :Tuple = 16 return intro_sort(_SCREAMING_SNAKE_CASE , 0 , len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(_SCREAMING_SNAKE_CASE ) max_depth -= 1 lowerCAmelCase__ :Any = median_of_a(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , start + ((end - start) // 2) + 1 , end - 1 ) lowerCAmelCase__ :Tuple = partition(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) intro_sort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :str = p return insertion_sort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() __A = input("""Enter numbers separated by a comma : """).strip() __A = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
93
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __UpperCAmelCase : str = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Any ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __UpperCAmelCase : Any = features.copy() if features else default_expected_features __UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=__lowerCamelCase , cache_dir=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: __UpperCAmelCase : Dict = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): __UpperCAmelCase : Optional[int] = tmp_path / """cache""" __UpperCAmelCase : str = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() __UpperCAmelCase : Optional[int] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Dict = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): __UpperCAmelCase : int = tmp_path / """cache""" __UpperCAmelCase : int = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Any = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Optional[int] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() with pytest.raises(__lowerCamelCase ): SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
63
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE = { 'configuration_pix2struct': [ 'PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Pix2StructConfig', 'Pix2StructTextConfig', 'Pix2StructVisionConfig', ], 'processing_pix2struct': ['Pix2StructProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ['Pix2StructImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ 'PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Pix2StructPreTrainedModel', 'Pix2StructForConditionalGeneration', 'Pix2StructVisionModel', 'Pix2StructTextModel', ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
94
from __future__ import annotations a : Optional[Any] = [True] * 1_000_001 a : Union[str, Any] = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): a : Optional[Any] = False i += 1 def lowerCamelCase__ ( __lowerCamelCase : int ): return seive[n] def lowerCamelCase__ ( __lowerCamelCase : int ): return any(digit in """02468""" for digit in str(__lowerCamelCase ) ) def lowerCamelCase__ ( __lowerCamelCase : int = 1000000 ): __UpperCAmelCase : Optional[Any] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__lowerCamelCase ) and not contains_an_even_digit(__lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) __UpperCAmelCase : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__lowerCamelCase ) )] if all(is_prime(__lowerCamelCase ) for i in list_nums ): result.append(__lowerCamelCase ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"""{len(find_circular_primes()) = }""")
63
0
"""simple docstring""" from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
95
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() a : Dict = logging.get_logger(__name__) a : Tuple = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowerCamelCase__ ( __lowerCamelCase : Dict ): if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase : Union[str, Any] = k.replace(__lowerCamelCase , __lowerCamelCase ) if k.startswith("""encoder""" ): __UpperCAmelCase : List[str] = k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase : Optional[Any] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : Union[str, Any] = k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase : Optional[int] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : List[Any] = k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase : Any = k.replace("""norm3""" , """final_layer_norm""" ) return k def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Optional[Any] = [ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase : Dict = sd.pop(__lowerCamelCase ) __UpperCAmelCase : List[str] = k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase : List[str] = v a : Optional[int] = ["START"] @torch.no_grad() def lowerCamelCase__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : str ): __UpperCAmelCase : str = torch.load(__lowerCamelCase , map_location="""cpu""" ) __UpperCAmelCase : Tuple = model["""model"""] __UpperCAmelCase : int = BlenderbotConfig.from_json_file(__lowerCamelCase ) __UpperCAmelCase : List[str] = BlenderbotForConditionalGeneration(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = m.model.state_dict().keys() __UpperCAmelCase : Any = [] __UpperCAmelCase : Any = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase : int = rename_state_dict_key(__lowerCamelCase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase : str = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__lowerCamelCase ) m.model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) m.half() m.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) a : Any = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
63
0
"""simple docstring""" import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class __A ( SCREAMING_SNAKE_CASE_ ): def __init__( self : int , __snake_case : UNetaDModel , __snake_case : UNetaDModel , __snake_case : DDPMScheduler , __snake_case : str , ) -> Tuple: super().__init__() __magic_name__: Dict = value_function __magic_name__: Optional[Any] = unet __magic_name__: Dict = scheduler __magic_name__: Optional[int] = env __magic_name__: List[Any] = env.get_dataset() __magic_name__: str = {} for key in self.data.keys(): try: __magic_name__: Any = self.data[key].mean() except: # noqa: E722 pass __magic_name__: Optional[int] = {} for key in self.data.keys(): try: __magic_name__: Dict = self.data[key].std() except: # noqa: E722 pass __magic_name__: List[str] = env.observation_space.shape[0] __magic_name__: Any = env.action_space.shape[0] def lowerCamelCase__ ( self : Any , __snake_case : Dict , __snake_case : int ) -> Optional[int]: return (x_in - self.means[key]) / self.stds[key] def lowerCamelCase__ ( self : List[Any] , __snake_case : Optional[Any] , __snake_case : List[Any] ) -> Optional[int]: return x_in * self.stds[key] + self.means[key] def lowerCamelCase__ ( self : List[Any] , __snake_case : List[Any] ) -> Tuple: if type(__snake_case ) is dict: return {k: self.to_torch(__snake_case ) for k, v in x_in.items()} elif torch.is_tensor(__snake_case ): return x_in.to(self.unet.device ) return torch.tensor(__snake_case , device=self.unet.device ) def lowerCamelCase__ ( self : str , __snake_case : str , __snake_case : Dict , __snake_case : Tuple ) -> Optional[Any]: for key, val in cond.items(): __magic_name__: int = val.clone() return x_in def lowerCamelCase__ ( self : Any , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : Any , __snake_case : List[Any] ) -> List[Any]: __magic_name__: str = x.shape[0] __magic_name__: Optional[Any] = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model __magic_name__: Dict = torch.full((batch_size,) , __snake_case , device=self.unet.device , dtype=torch.long ) for _ in range(__snake_case ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models __magic_name__: int = self.value_function(x.permute(0 , 2 , 1 ) , __snake_case ).sample __magic_name__: str = torch.autograd.grad([y.sum()] , [x] )[0] __magic_name__: Optional[int] = self.scheduler._get_variance(__snake_case ) __magic_name__: Optional[Any] = torch.exp(0.5 * posterior_variance ) __magic_name__: List[Any] = model_std * grad __magic_name__: List[str] = 0 __magic_name__: Any = x.detach() __magic_name__: Dict = x + scale * grad __magic_name__: Optional[Any] = self.reset_xa(__snake_case , __snake_case , self.action_dim ) __magic_name__: int = self.unet(x.permute(0 , 2 , 1 ) , __snake_case ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg __magic_name__: Dict = self.scheduler.step(__snake_case , __snake_case , __snake_case , predict_epsilon=__snake_case )["""prev_sample"""] # apply conditions to the trajectory (set the initial state) __magic_name__: Dict = self.reset_xa(__snake_case , __snake_case , self.action_dim ) __magic_name__: Union[str, Any] = self.to_torch(__snake_case ) return x, y def __call__( self : Optional[Any] , __snake_case : int , __snake_case : Tuple=6_4 , __snake_case : Union[str, Any]=3_2 , __snake_case : str=2 , __snake_case : Any=0.1 ) -> int: # normalize the observations and create batch dimension __magic_name__: List[str] = self.normalize(__snake_case , """observations""" ) __magic_name__: Optional[int] = obs[None].repeat(__snake_case , axis=0 ) __magic_name__: str = {0: self.to_torch(__snake_case )} __magic_name__: Any = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) __magic_name__: int = randn_tensor(__snake_case , device=self.unet.device ) __magic_name__: Union[str, Any] = self.reset_xa(__snake_case , __snake_case , self.action_dim ) __magic_name__: int = self.to_torch(__snake_case ) # run the diffusion process __magic_name__, __magic_name__: int = self.run_diffusion(__snake_case , __snake_case , __snake_case , __snake_case ) # sort output trajectories by value __magic_name__: List[str] = y.argsort(0 , descending=__snake_case ).squeeze() __magic_name__: List[str] = x[sorted_idx] __magic_name__: Dict = sorted_values[:, :, : self.action_dim] __magic_name__: str = actions.detach().cpu().numpy() __magic_name__: Optional[Any] = self.de_normalize(__snake_case , key="""actions""" ) # select the action with the highest value if y is not None: __magic_name__: Tuple = 0 else: # if we didn't run value guiding, select a random action __magic_name__: List[str] = np.random.randint(0 , __snake_case ) __magic_name__: int = denorm_actions[selected_index, 0] return denorm_actions
96
def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ): __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : List[str] = len(__lowerCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Union[str, Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None __UpperCAmelCase : str = sorted_collection[point] if current_item == item: return point else: if point < left: __UpperCAmelCase : Optional[Any] = left __UpperCAmelCase : Tuple = point elif point > right: __UpperCAmelCase : Optional[Any] = right __UpperCAmelCase : Dict = point else: if item < current_item: __UpperCAmelCase : Union[str, Any] = point - 1 else: __UpperCAmelCase : str = point + 1 return None def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) elif point > right: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , point - 1 ) else: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , point + 1 , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : int ): if collection != sorted(__lowerCamelCase ): raise ValueError("""Collection must be ascending sorted""" ) return True if __name__ == "__main__": import sys a : Optional[Any] = 0 if debug == 1: a : Optional[Any] = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") a : Tuple = 67 a : List[Any] = interpolation_search(collection, target) if result is not None: print(f"""{target} found at positions: {result}""") else: print("Not found")
63
0
from datetime import datetime as dt import os from github import Github __a = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def a ( ): '''simple docstring''' lowercase_ = Github(os.environ['''GITHUB_TOKEN'''] ) lowercase_ = g.get_repo('''huggingface/transformers''' ) lowercase_ = repo.get_issues(state='''open''' ) for issue in open_issues: lowercase_ = sorted([comment for comment in issue.get_comments()] , key=lambda snake_case__ : i.created_at , reverse=snake_case__ ) lowercase_ = comments[0] if len(snake_case__ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='''closed''' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) if __name__ == "__main__": main()
97
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
63
0
'''simple docstring''' from math import factorial lowercase__ : dict[str, int] = {str(digit): factorial(digit) for digit in range(10)} def a__ ( lowercase : int ) -> int: """simple docstring""" if not isinstance(lowercase, lowercase ): raise TypeError('''Parameter number must be int''' ) if number < 0: raise ValueError('''Parameter number must be greater than or equal to 0''' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(lowercase ) ) def a__ ( lowercase : int = 60, lowercase : int = 1000000 ) -> int: """simple docstring""" if not isinstance(lowercase, lowercase ) or not isinstance(lowercase, lowercase ): raise TypeError('''Parameters chain_length and number_limit must be int''' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( '''Parameters chain_length and number_limit must be greater than 0''' ) # the counter for the chains with the exact desired length _UpperCamelCase = 0 # the cached sizes of the previous chains _UpperCamelCase = {} for start_chain_element in range(1, lowercase ): # The temporary set will contain the elements of the chain _UpperCamelCase = set() _UpperCamelCase = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. _UpperCamelCase = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(lowercase ) chain_set_length += 1 _UpperCamelCase = digit_factorial_sum(lowercase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] _UpperCamelCase = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
98
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class a ( lowercase__ , lowercase__ ): """simple docstring""" a : Dict = 1 @register_to_config def __init__( self : int , __lowercase : int = 1000 , __lowercase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: # set `betas`, `alphas`, `timesteps` self.set_timesteps(__lowercase ) # standard deviation of the initial noise distribution __UpperCAmelCase : List[Any] = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __UpperCAmelCase : List[Any] = 4 # running values __UpperCAmelCase : str = [] def UpperCAmelCase ( self : Union[str, Any] , __lowercase : int , __lowercase : Union[str, torch.device] = None ) -> int: __UpperCAmelCase : int = num_inference_steps __UpperCAmelCase : Union[str, Any] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __UpperCAmelCase : Union[str, Any] = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __UpperCAmelCase : Dict = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __UpperCAmelCase : Dict = torch.sin(steps * math.pi / 2 ) ** 2 __UpperCAmelCase : List[Any] = (1.0 - self.betas**2) ** 0.5 __UpperCAmelCase : Tuple = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __UpperCAmelCase : Dict = timesteps.to(__lowercase ) __UpperCAmelCase : Optional[Any] = [] def UpperCAmelCase ( self : Optional[int] , __lowercase : torch.FloatTensor , __lowercase : int , __lowercase : torch.FloatTensor , __lowercase : bool = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __UpperCAmelCase : List[str] = (self.timesteps == timestep).nonzero().item() __UpperCAmelCase : Optional[Any] = timestep_index + 1 __UpperCAmelCase : List[str] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowercase ) if len(self.ets ) == 1: __UpperCAmelCase : Tuple = self.ets[-1] elif len(self.ets ) == 2: __UpperCAmelCase : Union[str, Any] = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __UpperCAmelCase : Union[str, Any] = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __UpperCAmelCase : List[Any] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __UpperCAmelCase : Union[str, Any] = self._get_prev_sample(__lowercase , __lowercase , __lowercase , __lowercase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowercase ) def UpperCAmelCase ( self : Optional[Any] , __lowercase : torch.FloatTensor , *__lowercase : Optional[Any] , **__lowercase : Any ) -> torch.FloatTensor: return sample def UpperCAmelCase ( self : Tuple , __lowercase : Tuple , __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Dict ) -> str: __UpperCAmelCase : int = self.alphas[timestep_index] __UpperCAmelCase : Tuple = self.betas[timestep_index] __UpperCAmelCase : Any = self.alphas[prev_timestep_index] __UpperCAmelCase : List[str] = self.betas[prev_timestep_index] __UpperCAmelCase : List[str] = (sample - sigma * ets) / max(__lowercase , 1e-8 ) __UpperCAmelCase : List[Any] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Tuple ) -> str: return self.config.num_train_timesteps
63
0
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() SCREAMING_SNAKE_CASE = { 'bart': ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'bert': ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-base-cased-finetuned-mrpc': ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'dpr': ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'gpt2': ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlnet': ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm': ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm-roberta': ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'transfo-xl': ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'openai-gpt': ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'roberta': ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'layoutlm': ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'roberta-large-mnli': ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'camembert': ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'flaubert': ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert': ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert-base-distilled-squad': ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert': ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert-visual-feature-encoder': ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'ctrl': ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'albert': ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 't5': ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'electra': ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'wav2vec2': ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) __a , __a , __a , __a = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: __a = cached_file(lowerCAmelCase__ , lowerCAmelCase__ , force_download=not use_cached_models ) __a = config_class.from_json_file(lowerCAmelCase__ ) __a = True __a = True print(f'''Building TensorFlow model from configuration: {config}''' ) __a = model_class(lowerCAmelCase__ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): __a = cached_file( lowerCAmelCase__ , lowerCAmelCase__ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: __a = load_pytorch_checkpoint_in_tfa_model(lowerCAmelCase__ , lowerCAmelCase__ ) if compare_with_pt_model: __a = tf_model(tf_model.dummy_inputs , training=lowerCAmelCase__ ) # build the network __a = torch.load(lowerCAmelCase__ , map_location="""cpu""" ) __a = pt_model_class.from_pretrained( pretrained_model_name_or_path=lowerCAmelCase__ , config=lowerCAmelCase__ , state_dict=lowerCAmelCase__ ) with torch.no_grad(): __a = pt_model(**pt_model.dummy_inputs ) __a = pto[0].numpy() __a = tfo[0].numpy() __a = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2E-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(lowerCAmelCase__ , save_format="""h5""" ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , ): if args_model_type is None: __a = list(MODEL_CLASSES.keys() ) else: __a = [args_model_type] for j, model_type in enumerate(lowerCAmelCase__ , start=1 ): print("""=""" * 100 ) print(f''' Converting model type {j}/{len(lowerCAmelCase__ )}: {model_type}''' ) print("""=""" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) __a , __a , __a , __a , __a = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: __a = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: __a = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(lowerCAmelCase__ , lowerCAmelCase__ ) , start=1 ): print("""-""" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue __a = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(lowerCAmelCase__ )}: {model_shortcut_name} - model_type {model_type}''' ) print("""-""" * 100 ) if config_shortcut_name in aws_config_map: __a = cached_file(lowerCAmelCase__ , lowerCAmelCase__ , force_download=not use_cached_models ) else: __a = config_shortcut_name if model_shortcut_name in aws_model_maps: __a = cached_file(lowerCAmelCase__ , lowerCAmelCase__ , force_download=not use_cached_models ) else: __a = model_shortcut_name if os.path.isfile(lowerCAmelCase__ ): __a = """converted_model""" convert_pt_checkpoint_to_tf( model_type=lowerCAmelCase__ , pytorch_checkpoint_path=lowerCAmelCase__ , config_file=lowerCAmelCase__ , tf_dump_path=os.path.join(lowerCAmelCase__ , model_shortcut_name + """-tf_model.h5""" ) , compare_with_pt_model=lowerCAmelCase__ , ) if remove_cached_files: os.remove(lowerCAmelCase__ ) os.remove(lowerCAmelCase__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_dump_path', default=None, type=str, required=True, help='Path to the output Tensorflow dump file.' ) parser.add_argument( '--model_type', default=None, type=str, help=( f'''Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ''' 'convert all the models from AWS.' ), ) parser.add_argument( '--pytorch_checkpoint_path', default=None, type=str, help=( 'Path to the PyTorch checkpoint path or shortcut name to download from AWS. ' 'If not given, will download and convert all the checkpoints from AWS.' ), ) parser.add_argument( '--config_file', default=None, type=str, help=( 'The config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture. If not given and ' '--pytorch_checkpoint_path is not given or is a shortcut name ' 'use the configuration associated to the shortcut name on the AWS' ), ) parser.add_argument( '--compare_with_pt_model', action='store_true', help='Compare Tensorflow and PyTorch model predictions.' ) parser.add_argument( '--use_cached_models', action='store_true', help='Use cached models if possible instead of updating to latest checkpoint versions.', ) parser.add_argument( '--remove_cached_files', action='store_true', help='Remove pytorch models after conversion (save memory when converting in batches).', ) parser.add_argument('--only_convert_finetuned_models', action='store_true', help='Only convert finetuned models.') SCREAMING_SNAKE_CASE = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
99
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCamelCase__ ( ): __UpperCAmelCase : Union[str, Any] = ArgumentParser("""Transformers CLI tool""" , usage="""transformers-cli <command> [<args>]""" ) __UpperCAmelCase : Any = parser.add_subparsers(help="""transformers-cli command helpers""" ) # Register commands ConvertCommand.register_subcommand(__lowerCamelCase ) DownloadCommand.register_subcommand(__lowerCamelCase ) EnvironmentCommand.register_subcommand(__lowerCamelCase ) RunCommand.register_subcommand(__lowerCamelCase ) ServeCommand.register_subcommand(__lowerCamelCase ) UserCommands.register_subcommand(__lowerCamelCase ) AddNewModelCommand.register_subcommand(__lowerCamelCase ) AddNewModelLikeCommand.register_subcommand(__lowerCamelCase ) LfsCommands.register_subcommand(__lowerCamelCase ) PTtoTFCommand.register_subcommand(__lowerCamelCase ) # Let's go __UpperCAmelCase : Optional[Any] = parser.parse_args() if not hasattr(__lowerCamelCase , """func""" ): parser.print_help() exit(1 ) # Run __UpperCAmelCase : Tuple = args.func(__lowerCamelCase ) service.run() if __name__ == "__main__": main()
63
0
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __snake_case ( lowerCAmelCase_ ) -> str: SCREAMING_SNAKE_CASE__ = [] for line in lines: SCREAMING_SNAKE_CASE__ = re.sub(r'''#.*''' , '''''' , lowerCAmelCase_ ) # remove comments if line: filtered_lines.append(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = '''\n'''.join(lowerCAmelCase_ ) # Make a hash from all this code SCREAMING_SNAKE_CASE__ = full_str.encode('''utf-8''' ) return shaaaa(lowerCAmelCase_ ).hexdigest() # get importable module names and hash for caching _A : str = { """csv""": (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), """json""": (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), """pandas""": (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), """parquet""": (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), """arrow""": (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), """text""": (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), """imagefolder""": (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), """audiofolder""": (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _A : Union[str, Any] = { """.csv""": ("""csv""", {}), """.tsv""": ("""csv""", {"""sep""": """\t"""}), """.json""": ("""json""", {}), """.jsonl""": ("""json""", {}), """.parquet""": ("""parquet""", {}), """.arrow""": ("""arrow""", {}), """.txt""": ("""text""", {}), } _EXTENSION_TO_MODULE.update({ext: ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _A : List[str] = {"""imagefolder""", """audiofolder"""} # Used to filter data files based on extensions given a module name _A : Dict[str, List[str]] = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append(""".zip""") _MODULE_TO_EXTENSIONS["audiofolder"].append(""".zip""")
100
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : List[str] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Dict = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
63
0
from collections.abc import Sequence from queue import Queue class __lowercase : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = start SCREAMING_SNAKE_CASE_ : Dict = end SCREAMING_SNAKE_CASE_ : Any = val SCREAMING_SNAKE_CASE_ : str = (start + end) // 2 SCREAMING_SNAKE_CASE_ : Union[str, Any] = left SCREAMING_SNAKE_CASE_ : Optional[int] = right def __repr__( self ): """simple docstring""" return F'''SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})''' class __lowercase : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = collection SCREAMING_SNAKE_CASE_ : List[str] = function if self.collection: SCREAMING_SNAKE_CASE_ : Union[str, Any] = self._build_tree(0 , len(lowerCAmelCase__ ) - 1 ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" self._update_tree(self.root , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" return self._query_range(self.root , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" if start == end: return SegmentTreeNode(lowerCAmelCase__ , lowerCAmelCase__ , self.collection[start] ) SCREAMING_SNAKE_CASE_ : Dict = (start + end) // 2 SCREAMING_SNAKE_CASE_ : List[Any] = self._build_tree(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = self._build_tree(mid + 1 , lowerCAmelCase__ ) return SegmentTreeNode(lowerCAmelCase__ , lowerCAmelCase__ , self.fn(left.val , right.val ) , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" if node.start == i and node.end == i: SCREAMING_SNAKE_CASE_ : List[str] = val return if i <= node.mid: self._update_tree(node.left , lowerCAmelCase__ , lowerCAmelCase__ ) else: self._update_tree(node.right , lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Dict = self.fn(node.left.val , node.right.val ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , lowerCAmelCase__ , lowerCAmelCase__ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , lowerCAmelCase__ , node.mid ) , self._query_range(node.right , node.mid + 1 , lowerCAmelCase__ ) , ) else: # range in right child tree return self._query_range(node.right , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" if self.root is not None: SCREAMING_SNAKE_CASE_ : Union[str, Any] = Queue() queue.put(self.root ) while not queue.empty(): SCREAMING_SNAKE_CASE_ : List[str] = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('*' * 50) lowerCAmelCase__ : Union[str, Any] =SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
101
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys a : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union __magic_name__ : Any = TypeVar("""T""") __magic_name__ : Optional[Any] = Union[List[T], Tuple[T, ...]] __magic_name__ : Optional[Any] = Union[T, List[T], Dict[str, T]] __magic_name__ : Dict = Union[str, bytes, os.PathLike]
102
def lowerCamelCase__ ( __lowerCamelCase : int ): if num <= 0: raise ValueError("""Input must be a positive integer""" ) __UpperCAmelCase : int = [True] * (num + 1) __UpperCAmelCase : Tuple = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCamelCase ): __UpperCAmelCase : str = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() a : Any = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
63
0
"""simple docstring""" from __future__ import annotations import os from typing import Any import requests snake_case = '''https://api.github.com''' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user snake_case = BASE_URL + '''/user''' # https://github.com/settings/tokens snake_case = os.environ.get('''USER_TOKEN''', '''''') def snake_case ( lowerCAmelCase_ ) -> dict[Any, Any]: _snake_case = { '''Authorization''': f"""token {auth_token}""", '''Accept''': '''application/vnd.github.v3+json''', } return requests.get(lowerCAmelCase_ , headers=lowerCAmelCase_ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F"{key}: {value}") else: raise ValueError('''\'USER_TOKEN\' field cannot be empty.''')
103
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Union[str, Any] = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class a ( lowercase__ ): """simple docstring""" a : Optional[int] = 'git_vision_model' def __init__( self : str , __lowercase : List[str]=768 , __lowercase : List[str]=3072 , __lowercase : List[Any]=12 , __lowercase : Dict=12 , __lowercase : int=3 , __lowercase : Any=224 , __lowercase : Optional[int]=16 , __lowercase : Dict="quick_gelu" , __lowercase : Any=1e-5 , __lowercase : str=0.0 , __lowercase : int=0.02 , **__lowercase : int , ) -> List[str]: super().__init__(**__lowercase ) __UpperCAmelCase : int = hidden_size __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Dict = num_attention_heads __UpperCAmelCase : int = num_channels __UpperCAmelCase : str = patch_size __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : int = initializer_range __UpperCAmelCase : Tuple = attention_dropout __UpperCAmelCase : Optional[int] = layer_norm_eps __UpperCAmelCase : List[Any] = hidden_act @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : Union[str, os.PathLike] , **__lowercase : List[str] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__lowercase ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("""model_type""" ) == "git": __UpperCAmelCase : str = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a ( lowercase__ ): """simple docstring""" a : List[str] = 'git' def __init__( self : Optional[int] , __lowercase : List[Any]=None , __lowercase : Tuple=30522 , __lowercase : str=768 , __lowercase : Optional[int]=6 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[int]=3072 , __lowercase : List[str]="gelu" , __lowercase : Tuple=0.1 , __lowercase : Optional[int]=0.1 , __lowercase : Optional[int]=1024 , __lowercase : Union[str, Any]=0.02 , __lowercase : Optional[Any]=1e-1_2 , __lowercase : List[Any]=0 , __lowercase : Dict="absolute" , __lowercase : Dict=True , __lowercase : Any=False , __lowercase : Optional[int]=101 , __lowercase : str=102 , __lowercase : Union[str, Any]=None , **__lowercase : Dict , ) -> Tuple: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: __UpperCAmelCase : Optional[int] = {} logger.info("""vision_config is None. initializing the GitVisionConfig with default values.""" ) __UpperCAmelCase : Tuple = GitVisionConfig(**__lowercase ) __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Union[str, Any] = hidden_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : List[str] = hidden_act __UpperCAmelCase : str = intermediate_size __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : int = max_position_embeddings __UpperCAmelCase : str = initializer_range __UpperCAmelCase : str = layer_norm_eps __UpperCAmelCase : Union[str, Any] = position_embedding_type __UpperCAmelCase : Dict = use_cache __UpperCAmelCase : int = tie_word_embeddings __UpperCAmelCase : Optional[int] = num_image_with_embedding __UpperCAmelCase : Optional[int] = bos_token_id __UpperCAmelCase : List[Any] = eos_token_id def UpperCAmelCase ( self : str ) -> int: __UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : List[str] = self.vision_config.to_dict() __UpperCAmelCase : Union[str, Any] = self.__class__.model_type return output
63
0
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
104
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = BarthezTokenizer a : Any = BarthezTokenizerFast a : Union[str, Any] = True a : Union[str, Any] = True def UpperCAmelCase ( self : Dict ) -> Any: super().setUp() __UpperCAmelCase : Optional[int] = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=__lowercase ) __UpperCAmelCase : str = tokenizer def UpperCAmelCase ( self : Optional[int] ) -> Tuple: __UpperCAmelCase : Dict = """<pad>""" __UpperCAmelCase : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> str: __UpperCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__lowercase ) , 101122 ) def UpperCAmelCase ( self : Any ) -> List[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __UpperCAmelCase : str = [0, 57, 3018, 70307, 91, 2] __UpperCAmelCase : List[Any] = self.tokenizer( __lowercase , max_length=len(__lowercase ) , padding=__lowercase , truncation=__lowercase , return_tensors="""pt""" ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __UpperCAmelCase : int = batch.input_ids.tolist()[0] self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> Tuple: if not self.test_rust_tokenizer: return __UpperCAmelCase : Union[str, Any] = self.get_tokenizer() __UpperCAmelCase : Optional[Any] = self.get_rust_tokenizer() __UpperCAmelCase : int = """I was born in 92000, and this is falsé.""" __UpperCAmelCase : Union[str, Any] = tokenizer.tokenize(__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = self.get_rust_tokenizer() __UpperCAmelCase : str = tokenizer.encode(__lowercase ) __UpperCAmelCase : Tuple = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: # fmt: off __UpperCAmelCase : str = {"""input_ids""": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __UpperCAmelCase : int = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=__lowercase , )
63
0
from PIL import Image def __UpperCAmelCase ( lowerCamelCase_ : Image ) -> Image: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[Any] = image.size SCREAMING_SNAKE_CASE_ : str = 0 SCREAMING_SNAKE_CASE_ : Dict = image.load() for i in range(lowerCamelCase_ ): for j in range(lowerCamelCase_ ): SCREAMING_SNAKE_CASE_ : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(lowerCamelCase_ ): for i in range(lowerCamelCase_ ): SCREAMING_SNAKE_CASE_ : List[Any] = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": UpperCamelCase__ : Any = mean_threshold(Image.open('''path_to_image''').convert('''L''')) image.save('''output_image_path''')
105
from __future__ import annotations import math def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : bool , __lowerCamelCase : list[int] , __lowerCamelCase : float ): if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(__lowerCamelCase ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = [90, 23, 6, 33, 21, 65, 123, 34423] __UpperCAmelCase : str = math.log(len(__lowerCamelCase ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase__ : def __UpperCamelCase ( self : int ) -> Dict: torch.manual_seed(0 ) A = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) A = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) A = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) A = DDPMScheduler( num_train_timesteps=1_000 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__UpperCamelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) A = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def __UpperCamelCase ( self : Any ) -> List[Any]: torch.manual_seed(0 ) A = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) A = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) A = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , class_embed_type='timestep' , mid_block_scale_factor=1.4_1_4 , time_embedding_act_fn='gelu' , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) A = DDPMScheduler( num_train_timesteps=1_000 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__UpperCamelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) A = DDPMScheduler( num_train_timesteps=1_000 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0_0_0_1 , beta_end=0.0_2 , ) torch.manual_seed(0 ) A = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def __UpperCamelCase ( self : List[Any] ) -> Dict: A = self.get_dummy_components() A = self.pipeline_class(**__UpperCamelCase ) pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) A = self.get_dummy_inputs(__UpperCamelCase ) A = inputs['prompt'] A = inputs['generator'] A = inputs['num_inference_steps'] A = inputs['output_type'] if "image" in inputs: A = inputs['image'] else: A = None if "mask_image" in inputs: A = inputs['mask_image'] else: A = None if "original_image" in inputs: A = inputs['original_image'] else: A = None A , A = pipe.encode_prompt(__UpperCamelCase ) # inputs with prompt converted to embeddings A = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: A = image if mask_image is not None: A = mask_image if original_image is not None: A = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) A = pipe(**__UpperCamelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__UpperCamelCase ) A = self.pipeline_class.from_pretrained(__UpperCamelCase ) pipe_loaded.to(__UpperCamelCase ) pipe_loaded.set_progress_bar_config(disable=__UpperCamelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__UpperCamelCase , __UpperCamelCase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) A = self.get_dummy_inputs(__UpperCamelCase ) A = inputs['generator'] A = inputs['num_inference_steps'] A = inputs['output_type'] # inputs with prompt converted to embeddings A = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: A = image if mask_image is not None: A = mask_image if original_image is not None: A = original_image A = pipe_loaded(**__UpperCamelCase )[0] A = np.abs(to_np(__UpperCamelCase ) - to_np(__UpperCamelCase ) ).max() self.assertLess(__UpperCamelCase , 1e-4 ) def __UpperCamelCase ( self : Optional[int] ) -> Dict: A = self.get_dummy_components() A = self.pipeline_class(**__UpperCamelCase ) pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) A = self.get_dummy_inputs(__UpperCamelCase ) A = pipe(**__UpperCamelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__UpperCamelCase ) A = self.pipeline_class.from_pretrained(__UpperCamelCase ) pipe_loaded.to(__UpperCamelCase ) pipe_loaded.set_progress_bar_config(disable=__UpperCamelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests A = self.get_dummy_inputs(__UpperCamelCase ) A = pipe_loaded(**__UpperCamelCase )[0] A = np.abs(to_np(__UpperCamelCase ) - to_np(__UpperCamelCase ) ).max() self.assertLess(__UpperCamelCase , 1e-4 )
106
from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class a ( lowercase__ ): """simple docstring""" a : Optional[Any] = 'openai-gpt' a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __lowercase : Tuple=40478 , __lowercase : Tuple=512 , __lowercase : int=768 , __lowercase : Dict=12 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[Any]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Optional[Any]=0.1 , __lowercase : Tuple=0.1 , __lowercase : Dict=1e-5 , __lowercase : Any=0.02 , __lowercase : List[str]="cls_index" , __lowercase : str=True , __lowercase : Dict=None , __lowercase : str=True , __lowercase : List[str]=0.1 , **__lowercase : List[Any] , ) -> List[Any]: __UpperCAmelCase : Optional[int] = vocab_size __UpperCAmelCase : Optional[Any] = n_positions __UpperCAmelCase : Optional[int] = n_embd __UpperCAmelCase : str = n_layer __UpperCAmelCase : Any = n_head __UpperCAmelCase : Tuple = afn __UpperCAmelCase : Any = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : str = attn_pdrop __UpperCAmelCase : str = layer_norm_epsilon __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : Optional[int] = summary_type __UpperCAmelCase : Optional[Any] = summary_use_proj __UpperCAmelCase : List[Any] = summary_activation __UpperCAmelCase : Union[str, Any] = summary_first_dropout __UpperCAmelCase : Dict = summary_proj_to_labels super().__init__(**__lowercase )
63
0
'''simple docstring''' _UpperCAmelCase : Any = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
107
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = KandinskyVaaInpaintPipeline a : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] a : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] a : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] a : List[Any] = False @property def UpperCAmelCase ( self : int ) -> Dict: return 32 @property def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: return 32 @property def UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def UpperCAmelCase ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return 100 @property def UpperCAmelCase ( self : Dict ) -> Any: torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } __UpperCAmelCase : int = UNetaDConditionModel(**__lowercase ) return model @property def UpperCAmelCase ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self : Any ) -> List[Any]: __UpperCAmelCase : List[str] = self.dummy_unet __UpperCAmelCase : List[str] = self.dummy_movq __UpperCAmelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowercase , ) __UpperCAmelCase : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCAmelCase ( self : str , __lowercase : Tuple , __lowercase : List[str]=0 ) -> Optional[Any]: __UpperCAmelCase : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(__lowercase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask __UpperCAmelCase : Union[str, Any] = np.ones((64, 64) , dtype=np.floataa ) __UpperCAmelCase : List[str] = 0 if str(__lowercase ).startswith("""mps""" ): __UpperCAmelCase : List[str] = torch.manual_seed(__lowercase ) else: __UpperCAmelCase : Optional[int] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __UpperCAmelCase : Optional[Any] = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = """cpu""" __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**__lowercase ) __UpperCAmelCase : Tuple = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(__lowercase ) ) __UpperCAmelCase : Tuple = output.images __UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] __UpperCAmelCase : str = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __UpperCAmelCase : Optional[Any] = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def UpperCAmelCase ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) __UpperCAmelCase : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) __UpperCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa ) __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = """a hat""" __UpperCAmelCase : str = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) __UpperCAmelCase : Any = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) __UpperCAmelCase : int = pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() __UpperCAmelCase : Optional[int] = pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
63
0
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict __a: Optional[Any] = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def _SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ) -> str: return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def _SCREAMING_SNAKE_CASE ( __snake_case ) -> List[Any]: _UpperCAmelCase = _TestCommandArgs(dataset=__snake_case , all_configs=__snake_case , save_infos=__snake_case ) _UpperCAmelCase = TestCommand(*__snake_case ) test_command.run() _UpperCAmelCase = os.path.join(__snake_case , """README.md""" ) assert os.path.exists(__snake_case ) _UpperCAmelCase = DatasetInfosDict.from_directory(__snake_case ) _UpperCAmelCase = DatasetInfosDict( { """default""": DatasetInfo( features=Features( { """tokens""": Sequence(Value("""string""" ) ), """ner_tags""": Sequence( ClassLabel(names=["""O""", """B-PER""", """I-PER""", """B-ORG""", """I-ORG""", """B-LOC""", """I-LOC"""] ) ), """langs""": Sequence(Value("""string""" ) ), """spans""": Sequence(Value("""string""" ) ), } ) , splits=[ { """name""": """train""", """num_bytes""": 2_3_5_1_5_6_3, """num_examples""": 1_0_0_0_0, }, { """name""": """validation""", """num_bytes""": 2_3_8_4_1_8, """num_examples""": 1_0_0_0, }, ] , download_size=3_9_4_0_6_8_0 , dataset_size=2_5_8_9_9_8_1 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: _UpperCAmelCase , _UpperCAmelCase = getattr(dataset_infos["""default"""] , __snake_case ), getattr(expected_dataset_infos["""default"""] , __snake_case ) if key == "num_bytes": assert is_apercent_close(__snake_case , __snake_case ) elif key == "splits": assert list(__snake_case ) == list(__snake_case ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
108
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a : List[Any] = True except ImportError: a : str = False try: from torch.hub import _get_torch_home a : List[Any] = _get_torch_home() except ImportError: a : int = os.path.expanduser( os.getenv("TORCH_HOME", os.path.join(os.getenv("XDG_CACHE_HOME", "~/.cache"), "torch")) ) a : Optional[Any] = os.path.join(torch_cache_home, "transformers") a : Optional[Any] = "https://cdn.huggingface.co" a : List[str] = "https://s3.amazonaws.com/models.huggingface.co/bert" a : Any = "/".join(str(Path(__file__).resolve()).split("/")[:-1]) a : Optional[int] = os.path.join(PATH, "config.yaml") a : Dict = os.path.join(PATH, "attributes.txt") a : Tuple = os.path.join(PATH, "objects.txt") a : Dict = os.getenv("PYTORCH_PRETRAINED_BERT_CACHE", default_cache_path) a : Dict = os.getenv("PYTORCH_TRANSFORMERS_CACHE", PYTORCH_PRETRAINED_BERT_CACHE) a : Optional[int] = os.getenv("TRANSFORMERS_CACHE", PYTORCH_TRANSFORMERS_CACHE) a : Any = "pytorch_model.bin" a : int = "config.yaml" def lowerCamelCase__ ( __lowerCamelCase : str=OBJECTS , __lowerCamelCase : Union[str, Any]=ATTRIBUTES ): __UpperCAmelCase : Union[str, Any] = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(""",""" )[0].lower().strip() ) __UpperCAmelCase : Dict = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(""",""" )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : List[str] = OrderedDict() with open(__lowerCamelCase , """rb""" ) as f: __UpperCAmelCase : int = pkl.load(__lowerCamelCase )["""model"""] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : List[Any] = ckp.pop(__lowerCamelCase ) if isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : Union[str, Any] = torch.tensor(__lowerCamelCase ) else: assert isinstance(__lowerCamelCase , torch.tensor ), type(__lowerCamelCase ) __UpperCAmelCase : List[str] = v return r class a : """simple docstring""" a : Dict = {} def __init__( self : Dict , __lowercase : dict , __lowercase : str = "root" , __lowercase : Any=0 ) -> Dict: __UpperCAmelCase : List[str] = name __UpperCAmelCase : str = level __UpperCAmelCase : int = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : List[str] = copy.deepcopy(__lowercase ) __UpperCAmelCase : Dict = copy.deepcopy(__lowercase ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : Union[str, Any] = Config(__lowercase , name=__lowercase , level=level + 1 ) __UpperCAmelCase : Union[str, Any] = v setattr(self , __lowercase , __lowercase ) __UpperCAmelCase : Any = d def __repr__( self : Optional[Any] ) -> Optional[int]: return str(list((self._pointer.keys()) ) ) def __setattr__( self : List[str] , __lowercase : List[str] , __lowercase : Tuple ) -> int: __UpperCAmelCase : int = val __UpperCAmelCase : List[str] = val __UpperCAmelCase : Union[str, Any] = key.split(""".""" ) __UpperCAmelCase : List[Any] = len(__lowercase ) - 1 __UpperCAmelCase : List[Any] = self._pointer if len(__lowercase ) > 1: for i, l in enumerate(__lowercase ): if hasattr(self , __lowercase ) and isinstance(getattr(self , __lowercase ) , __lowercase ): setattr(getattr(self , __lowercase ) , """.""".join(levels[i:] ) , __lowercase ) if l == last_level: __UpperCAmelCase : Union[str, Any] = val else: __UpperCAmelCase : Union[str, Any] = pointer[l] def UpperCAmelCase ( self : Tuple ) -> Optional[int]: return self._pointer def UpperCAmelCase ( self : str , __lowercase : Optional[int] , __lowercase : Any ) -> Optional[int]: with open(f"""{file_name}""" , """w""" ) as stream: dump(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] ) -> Any: with open(f"""{file_name}""" , """w""" ) as stream: json.dump(__lowercase , __lowercase ) @staticmethod def UpperCAmelCase ( __lowercase : List[Any] ) -> Optional[Any]: with open(__lowercase ) as stream: __UpperCAmelCase : Any = load(__lowercase , Loader=__lowercase ) return data def __str__( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = """ """ if self._name != "root": __UpperCAmelCase : Optional[Any] = f"""{t * (self._level-1)}{self._name}:\n""" else: __UpperCAmelCase : List[Any] = """""" __UpperCAmelCase : Optional[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__lowercase , __lowercase ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__lowercase ).__name__})\n""" __UpperCAmelCase : int = level return r[:-1] @classmethod def UpperCAmelCase ( cls : List[str] , __lowercase : str , **__lowercase : Any ) -> Any: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) return cls(__lowercase ) @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : str , **__lowercase : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase : int = kwargs.pop("""cache_dir""" , __lowercase ) __UpperCAmelCase : int = kwargs.pop("""force_download""" , __lowercase ) __UpperCAmelCase : str = kwargs.pop("""resume_download""" , __lowercase ) __UpperCAmelCase : Dict = kwargs.pop("""proxies""" , __lowercase ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""local_files_only""" , __lowercase ) if os.path.isdir(__lowercase ): __UpperCAmelCase : List[Any] = os.path.join(__lowercase , __lowercase ) elif os.path.isfile(__lowercase ) or is_remote_url(__lowercase ): __UpperCAmelCase : Tuple = pretrained_model_name_or_path else: __UpperCAmelCase : Optional[int] = hf_bucket_url(__lowercase , filename=__lowercase , use_cdn=__lowercase ) try: # Load from URL or cache if already cached __UpperCAmelCase : Optional[int] = cached_path( __lowercase , cache_dir=__lowercase , force_download=__lowercase , proxies=__lowercase , resume_download=__lowercase , local_files_only=__lowercase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : Optional[int] = Config.load_yaml(__lowercase ) except EnvironmentError: __UpperCAmelCase : str = """Can't load config for""" raise EnvironmentError(__lowercase ) if resolved_config_file == config_file: print("""loading configuration file from path""" ) else: print("""loading configuration file cache""" ) return Config.load_yaml(__lowercase ), kwargs def lowerCamelCase__ ( __lowerCamelCase : Dict ): __UpperCAmelCase : Optional[int] = torch.load("""dump.pt""" , map_location=in_tensor.device ) __UpperCAmelCase : Tuple = in_tensor.numpy() __UpperCAmelCase : Optional[int] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception("""tensors are all good""" ) # Hugging face functions below def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Tuple = urlparse(__lowerCamelCase ) return parsed.scheme in ("http", "https") def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : int=True ): __UpperCAmelCase : int = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : Optional[int] = """/""" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=None , ): __UpperCAmelCase : Optional[int] = """python/{}""".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + "; ".join("""{}/{}""".format(__lowerCamelCase , __lowerCamelCase ) for k, v in user_agent.items() ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + user_agent __UpperCAmelCase : List[str] = {"""user-agent""": ua} if resume_size > 0: __UpperCAmelCase : Union[str, Any] = """bytes=%d-""" % (resume_size,) __UpperCAmelCase : Union[str, Any] = requests.get(__lowerCamelCase , stream=__lowerCamelCase , proxies=__lowerCamelCase , headers=__lowerCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : List[str] = response.headers.get("""Content-Length""" ) __UpperCAmelCase : str = resume_size + int(__lowerCamelCase ) if content_length is not None else None __UpperCAmelCase : List[Any] = tqdm( unit="""B""" , unit_scale=__lowerCamelCase , total=__lowerCamelCase , initial=__lowerCamelCase , desc="""Downloading""" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(__lowerCamelCase ) ) temp_file.write(__lowerCamelCase ) progress.close() def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=10 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=None , __lowerCamelCase : List[str]=False , ): if cache_dir is None: __UpperCAmelCase : Optional[Any] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[str] = str(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) __UpperCAmelCase : List[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(__lowerCamelCase , allow_redirects=__lowerCamelCase , proxies=__lowerCamelCase , timeout=__lowerCamelCase ) if response.status_code == 200: __UpperCAmelCase : Dict = response.headers.get("""ETag""" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : List[str] = url_to_filename(__lowerCamelCase , __lowerCamelCase ) # get cache path to put the file __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , __lowerCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(__lowerCamelCase ): return cache_path else: __UpperCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(__lowerCamelCase ) , filename + """.*""" ) if not file.endswith(""".json""" ) and not file.endswith(""".lock""" ) ] if len(__lowerCamelCase ) > 0: return os.path.join(__lowerCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( """Cannot find the requested files in the cached path and outgoing traffic has been""" """ disabled. To enable model look-ups and downloads online, set 'local_files_only'""" """ to False.""" ) return None # From now on, etag is not None. if os.path.exists(__lowerCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : str = cache_path + """.lock""" with FileLock(__lowerCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(__lowerCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : int = cache_path + """.incomplete""" @contextmanager def _resumable_file_manager(): with open(__lowerCamelCase , """a+b""" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(__lowerCamelCase ): __UpperCAmelCase : List[Any] = os.stat(__lowerCamelCase ).st_size else: __UpperCAmelCase : List[Any] = 0 else: __UpperCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=__lowerCamelCase , delete=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( """%s not found in cache or force_download set to True, downloading to %s""" , __lowerCamelCase , temp_file.name , ) http_get( __lowerCamelCase , __lowerCamelCase , proxies=__lowerCamelCase , resume_size=__lowerCamelCase , user_agent=__lowerCamelCase , ) os.replace(temp_file.name , __lowerCamelCase ) __UpperCAmelCase : Any = {"""url""": url, """etag""": etag} __UpperCAmelCase : Union[str, Any] = cache_path + """.json""" with open(__lowerCamelCase , """w""" ) as meta_file: json.dump(__lowerCamelCase , __lowerCamelCase ) return cache_path def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any]=None ): __UpperCAmelCase : Tuple = url.encode("""utf-8""" ) __UpperCAmelCase : Optional[Any] = shaaaa(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = url_hash.hexdigest() if etag: __UpperCAmelCase : int = etag.encode("""utf-8""" ) __UpperCAmelCase : List[str] = shaaaa(__lowerCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(""".h5""" ): filename += ".h5" return filename def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : int=None , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=False , ): if cache_dir is None: __UpperCAmelCase : List[str] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Any = str(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) if is_remote_url(__lowerCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : Tuple = get_from_cache( __lowerCamelCase , cache_dir=__lowerCamelCase , force_download=__lowerCamelCase , proxies=__lowerCamelCase , resume_download=__lowerCamelCase , user_agent=__lowerCamelCase , local_files_only=__lowerCamelCase , ) elif os.path.exists(__lowerCamelCase ): # File, and it exists. __UpperCAmelCase : Tuple = url_or_filename elif urlparse(__lowerCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("""file {} not found""".format(__lowerCamelCase ) ) else: # Something unknown raise ValueError("""unable to parse {} as a URL or as a local path""".format(__lowerCamelCase ) ) if extract_compressed_file: if not is_zipfile(__lowerCamelCase ) and not tarfile.is_tarfile(__lowerCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : int = os.path.split(__lowerCamelCase ) __UpperCAmelCase : Any = output_file.replace(""".""" , """-""" ) + """-extracted""" __UpperCAmelCase : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isdir(__lowerCamelCase ) and os.listdir(__lowerCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : str = output_path + """.lock""" with FileLock(__lowerCamelCase ): shutil.rmtree(__lowerCamelCase , ignore_errors=__lowerCamelCase ) os.makedirs(__lowerCamelCase ) if is_zipfile(__lowerCamelCase ): with ZipFile(__lowerCamelCase , """r""" ) as zip_file: zip_file.extractall(__lowerCamelCase ) zip_file.close() elif tarfile.is_tarfile(__lowerCamelCase ): __UpperCAmelCase : Any = tarfile.open(__lowerCamelCase ) tar_file.extractall(__lowerCamelCase ) tar_file.close() else: raise EnvironmentError("""Archive format of {} could not be identified""".format(__lowerCamelCase ) ) return output_path_extracted return output_path def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : int="," ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): with open(__lowerCamelCase ) as f: __UpperCAmelCase : List[Any] = eval(f.read() ) else: __UpperCAmelCase : List[str] = requests.get(__lowerCamelCase ) try: __UpperCAmelCase : int = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : str = eval(__lowerCamelCase ) except Exception: __UpperCAmelCase : List[Any] = data.split("""\n""" ) req.close() return data def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : Optional[int] = requests.get(__lowerCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCamelCase__ ( __lowerCamelCase : str ): __UpperCAmelCase : int = url.split("""/""" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(__lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as stream: __UpperCAmelCase : List[str] = pkl.load(__lowerCamelCase ) __UpperCAmelCase : Dict = weights.pop("""model""" ) __UpperCAmelCase : Union[str, Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(__lowerCamelCase ) if "running_var" in k: __UpperCAmelCase : Optional[int] = torch.tensor([0] ) __UpperCAmelCase : Tuple = k.replace("""running_var""" , """num_batches_tracked""" ) __UpperCAmelCase : Any = zero return new def lowerCamelCase__ ( ): print(f"""{os.path.abspath(os.path.join(__lowerCamelCase , os.pardir ) )}/demo.ipynb""" ) def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]="RGB" ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): __UpperCAmelCase : List[str] = cva.imread(__lowerCamelCase ) else: __UpperCAmelCase : int = get_image_from_url(__lowerCamelCase ) assert img is not None, f"""could not connect to: {im}""" __UpperCAmelCase : Any = cva.cvtColor(__lowerCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Optional[int] = img[:, :, ::-1] return img def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int=1 ): return (images[i : i + batch] for i in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ))
63
0
'''simple docstring''' import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase="pt" ) -> List[Any]: '''simple docstring''' __SCREAMING_SNAKE_CASE = {"""add_prefix_space""": True} if isinstance(__UpperCAmelCase , __UpperCAmelCase ) and not line.startswith(""" """ ) else {} __SCREAMING_SNAKE_CASE = padding_side return tokenizer( [line] , max_length=__UpperCAmelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCAmelCase , return_tensors=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , **__UpperCAmelCase , ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , ) -> Tuple: '''simple docstring''' __SCREAMING_SNAKE_CASE = input_ids.ne(__UpperCAmelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __a ( _snake_case ): def __init__( self : List[str] ,lowerCamelCase : Dict ,lowerCamelCase : Tuple ,lowerCamelCase : List[str] ,lowerCamelCase : Tuple ,lowerCamelCase : Tuple="train" ,lowerCamelCase : str=None ,lowerCamelCase : Union[str, Any]=None ,lowerCamelCase : str=None ,lowerCamelCase : Optional[int]="" ,): '''simple docstring''' super().__init__() __SCREAMING_SNAKE_CASE = Path(lowerCamelCase ).joinpath(type_path + """.source""" ) __SCREAMING_SNAKE_CASE = Path(lowerCamelCase ).joinpath(type_path + """.target""" ) __SCREAMING_SNAKE_CASE = self.get_char_lens(self.src_file ) __SCREAMING_SNAKE_CASE = max_source_length __SCREAMING_SNAKE_CASE = max_target_length assert min(self.src_lens ) > 0, f"""found empty line in {self.src_file}""" __SCREAMING_SNAKE_CASE = tokenizer __SCREAMING_SNAKE_CASE = prefix if n_obs is not None: __SCREAMING_SNAKE_CASE = self.src_lens[:n_obs] __SCREAMING_SNAKE_CASE = src_lang __SCREAMING_SNAKE_CASE = tgt_lang def __len__( self : Optional[int] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,lowerCamelCase : Any ): '''simple docstring''' __SCREAMING_SNAKE_CASE = index + 1 # linecache starts at 1 __SCREAMING_SNAKE_CASE = self.prefix + linecache.getline(str(self.src_file ) ,lowerCamelCase ).rstrip("""\n""" ) __SCREAMING_SNAKE_CASE = linecache.getline(str(self.tgt_file ) ,lowerCamelCase ).rstrip("""\n""" ) assert source_line, f"""empty source line for index {index}""" assert tgt_line, f"""empty tgt line for index {index}""" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,lowerCamelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right __SCREAMING_SNAKE_CASE = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,lowerCamelCase ) else self.tokenizer ) __SCREAMING_SNAKE_CASE = self.tokenizer.generator if isinstance(self.tokenizer ,lowerCamelCase ) else self.tokenizer __SCREAMING_SNAKE_CASE = encode_line(lowerCamelCase ,lowerCamelCase ,self.max_source_length ,"""right""" ) __SCREAMING_SNAKE_CASE = encode_line(lowerCamelCase ,lowerCamelCase ,self.max_target_length ,"""right""" ) __SCREAMING_SNAKE_CASE = source_inputs["""input_ids"""].squeeze() __SCREAMING_SNAKE_CASE = target_inputs["""input_ids"""].squeeze() __SCREAMING_SNAKE_CASE = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def UpperCAmelCase__ ( lowerCamelCase : List[str] ): '''simple docstring''' return [len(lowerCamelCase ) for x in Path(lowerCamelCase ).open().readlines()] def UpperCAmelCase__ ( self : int ,lowerCamelCase : Union[str, Any] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = torch.stack([x["""input_ids"""] for x in batch] ) __SCREAMING_SNAKE_CASE = torch.stack([x["""attention_mask"""] for x in batch] ) __SCREAMING_SNAKE_CASE = torch.stack([x["""decoder_input_ids"""] for x in batch] ) __SCREAMING_SNAKE_CASE = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,lowerCamelCase ) else self.tokenizer.pad_token_id ) __SCREAMING_SNAKE_CASE = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,lowerCamelCase ) else self.tokenizer.pad_token_id ) __SCREAMING_SNAKE_CASE = trim_batch(lowerCamelCase ,lowerCamelCase ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = trim_batch(lowerCamelCase ,lowerCamelCase ,attention_mask=lowerCamelCase ) __SCREAMING_SNAKE_CASE = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch a = getLogger(__name__) def __magic_name__ ( __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCAmelCase ) ) def __magic_name__ ( __UpperCAmelCase ) -> None: '''simple docstring''' __SCREAMING_SNAKE_CASE = get_git_info() save_json(__UpperCAmelCase , os.path.join(__UpperCAmelCase , """git_log.json""" ) ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=4 , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' with open(__UpperCAmelCase , """w""" ) as f: json.dump(__UpperCAmelCase , __UpperCAmelCase , indent=__UpperCAmelCase , **__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Dict: '''simple docstring''' with open(__UpperCAmelCase ) as f: return json.load(__UpperCAmelCase ) def __magic_name__ ( ) -> List[str]: '''simple docstring''' __SCREAMING_SNAKE_CASE = git.Repo(search_parent_directories=__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = { """repo_id""": str(__UpperCAmelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> List: '''simple docstring''' return list(map(__UpperCAmelCase , __UpperCAmelCase ) ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' with open(__UpperCAmelCase , """wb""" ) as f: return pickle.dump(__UpperCAmelCase , __UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> str: '''simple docstring''' def remove_articles(__UpperCAmelCase ): return re.sub(R"""\b(a|an|the)\b""" , """ """ , __UpperCAmelCase ) def white_space_fix(__UpperCAmelCase ): return " ".join(text.split() ) def remove_punc(__UpperCAmelCase ): __SCREAMING_SNAKE_CASE = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCAmelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCAmelCase ) ) ) ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' __SCREAMING_SNAKE_CASE = normalize_answer(__UpperCAmelCase ).split() __SCREAMING_SNAKE_CASE = normalize_answer(__UpperCAmelCase ).split() __SCREAMING_SNAKE_CASE = Counter(__UpperCAmelCase ) & Counter(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = sum(common.values() ) if num_same == 0: return 0 __SCREAMING_SNAKE_CASE = 1.0 * num_same / len(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = 1.0 * num_same / len(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = (2 * precision * recall) / (precision + recall) return fa def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' return normalize_answer(__UpperCAmelCase ) == normalize_answer(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: '''simple docstring''' assert len(__UpperCAmelCase ) == len(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = 0 for hypo, pred in zip(__UpperCAmelCase , __UpperCAmelCase ): em += exact_match_score(__UpperCAmelCase , __UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: em /= len(__UpperCAmelCase ) return {"em": em} def __magic_name__ ( __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' return model_prefix.startswith("""rag""" ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' __SCREAMING_SNAKE_CASE = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead __SCREAMING_SNAKE_CASE = """dropout_rate""" for p in extra_params: if getattr(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): if not hasattr(__UpperCAmelCase , __UpperCAmelCase ) and not hasattr(__UpperCAmelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCAmelCase ) ) delattr(__UpperCAmelCase , __UpperCAmelCase ) continue __SCREAMING_SNAKE_CASE = p if hasattr(__UpperCAmelCase , __UpperCAmelCase ) else equivalent_param[p] setattr(__UpperCAmelCase , __UpperCAmelCase , getattr(__UpperCAmelCase , __UpperCAmelCase ) ) delattr(__UpperCAmelCase , __UpperCAmelCase ) return hparams, config
109
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Any=13 , __lowercase : Optional[int]=7 , __lowercase : str=True , __lowercase : Optional[Any]=True , __lowercase : int=True , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : int=5 , __lowercase : Tuple=4 , __lowercase : str=37 , __lowercase : Optional[int]="gelu" , __lowercase : Tuple=0.1 , __lowercase : str=0.1 , __lowercase : Dict=512 , __lowercase : List[Any]=16 , __lowercase : Dict=2 , __lowercase : Union[str, Any]=0.02 , __lowercase : Dict=4 , ) -> int: __UpperCAmelCase : Dict = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : str = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Dict = use_token_type_ids __UpperCAmelCase : Optional[int] = use_labels __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : str = type_sequence_label_size __UpperCAmelCase : List[str] = initializer_range __UpperCAmelCase : Tuple = num_choices def UpperCAmelCase ( self : Dict ) -> Tuple: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : List[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase ( self : Tuple ) -> List[Any]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase ( self : Any ) -> List[str]: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[int] = config_and_inputs __UpperCAmelCase : int = True __UpperCAmelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = True a : List[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : List[str] = FlaxRobertaModelTester(self ) @slow def UpperCAmelCase ( self : str ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase : Union[str, Any] = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowercase ) __UpperCAmelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowercase )
63
0
"""simple docstring""" import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class a ( lowercase ): UpperCamelCase : Tuple = """segformer""" def __init__( self , UpperCamelCase_=3 , UpperCamelCase_=4 , UpperCamelCase_=[2, 2, 2, 2] , UpperCamelCase_=[8, 4, 2, 1] , UpperCamelCase_=[32, 64, 160, 256] , UpperCamelCase_=[7, 3, 3, 3] , UpperCamelCase_=[4, 2, 2, 2] , UpperCamelCase_=[1, 2, 5, 8] , UpperCamelCase_=[4, 4, 4, 4] , UpperCamelCase_="gelu" , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.1 , UpperCamelCase_=0.02 , UpperCamelCase_=0.1 , UpperCamelCase_=1E-6 , UpperCamelCase_=256 , UpperCamelCase_=255 , **UpperCamelCase_ , ): super().__init__(**UpperCamelCase_ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( 'Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be' ' removed, as the behaviour will default to that of reshape_last_stage = True.' , UpperCamelCase_ , ) UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Dict = num_encoder_blocks UpperCAmelCase__ : List[Any] = depths UpperCAmelCase__ : Dict = sr_ratios UpperCAmelCase__ : Optional[int] = hidden_sizes UpperCAmelCase__ : Union[str, Any] = patch_sizes UpperCAmelCase__ : Optional[Any] = strides UpperCAmelCase__ : Tuple = mlp_ratios UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : Dict = hidden_dropout_prob UpperCAmelCase__ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase__ : Dict = classifier_dropout_prob UpperCAmelCase__ : Any = initializer_range UpperCAmelCase__ : Union[str, Any] = drop_path_rate UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : str = decoder_hidden_size UpperCAmelCase__ : Union[str, Any] = kwargs.get('reshape_last_stage' , UpperCamelCase_ ) UpperCAmelCase__ : Any = semantic_loss_ignore_index class a ( lowercase ): UpperCamelCase : List[str] = version.parse("""1.11""" ) @property def __snake_case ( self ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def __snake_case ( self ): return 1E-4 @property def __snake_case ( self ): return 12
110
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging a : Optional[int] = logging.get_logger(__name__) class a ( lowercase__ ): """simple docstring""" a : Tuple = 'linear' a : int = 'cosine' a : Optional[Any] = 'cosine_with_restarts' a : Dict = 'polynomial' a : Tuple = 'constant' a : Dict = 'constant_with_warmup' a : Any = 'piecewise_constant' def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int = -1 ): return LambdaLR(__lowerCamelCase , lambda __lowerCamelCase : 1 , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : str , __lowerCamelCase : int = -1 ): __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Tuple = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __UpperCAmelCase , __UpperCAmelCase : List[str] = rule_str.split(""":""" ) __UpperCAmelCase : Any = int(__lowerCamelCase ) __UpperCAmelCase : List[str] = float(__lowerCamelCase ) __UpperCAmelCase : int = value __UpperCAmelCase : Any = float(rule_list[-1] ) def create_rules_function(__lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): def rule_func(__lowerCamelCase : int ) -> float: __UpperCAmelCase : Tuple = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=-1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float = 0.5 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Dict ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Tuple = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int = 1 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Union[str, Any] ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Union[str, Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=1E-7 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : int=-1 ): __UpperCAmelCase : Tuple = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(f"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __UpperCAmelCase : Optional[Any] = lr_init - lr_end __UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps __UpperCAmelCase : int = 1 - (current_step - num_warmup_steps) / decay_steps __UpperCAmelCase : Optional[int] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a : int = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCamelCase__ ( __lowerCamelCase : Union[str, SchedulerType] , __lowerCamelCase : Optimizer , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 1 , __lowerCamelCase : float = 1.0 , __lowerCamelCase : int = -1 , ): __UpperCAmelCase : Union[str, Any] = SchedulerType(__lowerCamelCase ) __UpperCAmelCase : int = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
63
0
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
146
from math import pi, sqrt def lowerCamelCase__ ( __lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCamelCase__ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() a : Optional[int] = 1.0 while num: a : List[str] = float(input("Gamma of: ")) print(f"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
63
0
'''simple docstring''' import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ (lowercase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = LEDTokenizer SCREAMING_SNAKE_CASE : Optional[int] = LEDTokenizerFast SCREAMING_SNAKE_CASE : List[str] = True def SCREAMING_SNAKE_CASE ( self : Tuple ): super().setUp() __lowercase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] __lowercase = dict(zip(__lowercase ,range(len(__lowercase ) ) ) ) __lowercase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __lowercase = {"""unk_token""": """<unk>"""} __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowercase ) + '''\n''' ) with open(self.merges_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowercase ) ) def SCREAMING_SNAKE_CASE ( self : Any ,**lowercase__ : Dict ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**__lowercase ) def SCREAMING_SNAKE_CASE ( self : List[str] ,**lowercase__ : Any ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**__lowercase ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[Any] ): return "lower newer", "lower newer" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return LEDTokenizer.from_pretrained('''allenai/led-base-16384''' ) @cached_property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return LEDTokenizerFast.from_pretrained('''allenai/led-base-16384''' ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __lowercase = [0, 2_5_0, 2_5_1, 1_7_8_1_8, 1_3, 3_9_1_8_6, 1_9_3_8, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = tokenizer(__lowercase ,max_length=len(__lowercase ) ,padding=__lowercase ,return_tensors='''pt''' ) self.assertIsInstance(__lowercase ,__lowercase ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) __lowercase = batch.input_ids.tolist()[0] self.assertListEqual(__lowercase ,__lowercase ) @require_torch def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = tokenizer(__lowercase ,padding=__lowercase ,return_tensors='''pt''' ) self.assertIn('''input_ids''' ,__lowercase ) self.assertIn('''attention_mask''' ,__lowercase ) self.assertNotIn('''labels''' ,__lowercase ) self.assertNotIn('''decoder_attention_mask''' ,__lowercase ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = tokenizer(text_target=__lowercase ,max_length=3_2 ,padding='''max_length''' ,return_tensors='''pt''' ) self.assertEqual(3_2 ,targets['''input_ids'''].shape[1] ) @require_torch def SCREAMING_SNAKE_CASE ( self : str ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = tokenizer( ['''I am a small frog''' * 1_0_2_4, '''I am a small frog'''] ,padding=__lowercase ,truncation=__lowercase ,return_tensors='''pt''' ) self.assertIsInstance(__lowercase ,__lowercase ) self.assertEqual(batch.input_ids.shape ,(2, 5_1_2_2) ) @require_torch def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ["""A long paragraph for summarization."""] __lowercase = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = tokenizer(__lowercase ,return_tensors='''pt''' ) __lowercase = tokenizer(text_target=__lowercase ,return_tensors='''pt''' ) __lowercase = inputs["""input_ids"""] __lowercase = targets["""input_ids"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def SCREAMING_SNAKE_CASE ( self : int ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __lowercase = ["""Summary of the text.""", """Another summary."""] __lowercase = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] __lowercase = tokenizer(__lowercase ,padding=__lowercase ) __lowercase = [[0] * len(__lowercase ) for x in encoded_output["""input_ids"""]] __lowercase = tokenizer.pad(__lowercase ) self.assertSequenceEqual(outputs['''global_attention_mask'''] ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass def SCREAMING_SNAKE_CASE ( self : Tuple ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(__lowercase ,**__lowercase ) __lowercase = self.tokenizer_class.from_pretrained(__lowercase ,**__lowercase ) __lowercase = """A, <mask> AllenNLP sentence.""" __lowercase = tokenizer_r.encode_plus(__lowercase ,add_special_tokens=__lowercase ,return_token_type_ids=__lowercase ) __lowercase = tokenizer_p.encode_plus(__lowercase ,add_special_tokens=__lowercase ,return_token_type_ids=__lowercase ) self.assertEqual(sum(tokens_r['''token_type_ids'''] ) ,sum(tokens_p['''token_type_ids'''] ) ) self.assertEqual( sum(tokens_r['''attention_mask'''] ) / len(tokens_r['''attention_mask'''] ) ,sum(tokens_p['''attention_mask'''] ) / len(tokens_p['''attention_mask'''] ) ,) __lowercase = tokenizer_r.convert_ids_to_tokens(tokens_r['''input_ids'''] ) __lowercase = tokenizer_p.convert_ids_to_tokens(tokens_p['''input_ids'''] ) self.assertSequenceEqual(tokens_p['''input_ids'''] ,[0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual(tokens_r['''input_ids'''] ,[0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual( __lowercase ,['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) self.assertSequenceEqual( __lowercase ,['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] )
41
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class a : """simple docstring""" a : int a : Node | None = None a : Node | None = None def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = Node(1 ) __UpperCAmelCase : int = Node(2 ) __UpperCAmelCase : Optional[Any] = Node(3 ) __UpperCAmelCase : Dict = Node(4 ) __UpperCAmelCase : Tuple = Node(5 ) return tree def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCamelCase__ ( __lowerCamelCase : Node | None ): __UpperCAmelCase : list[Any] = [] if root is None: return output __UpperCAmelCase : Tuple = deque([root] ) while process_queue: __UpperCAmelCase : Optional[Any] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None ): if root is None: return [] __UpperCAmelCase : list[Sequence[Node | None]] = [] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : int = height(__lowerCamelCase ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : Optional[int] = 0 return output def lowerCamelCase__ ( ): # Main function for testing. __UpperCAmelCase : List[Any] = make_tree() print(f"""In-order Traversal: {inorder(__lowerCamelCase )}""" ) print(f"""Pre-order Traversal: {preorder(__lowerCamelCase )}""" ) print(f"""Post-order Traversal: {postorder(__lowerCamelCase )}""" , """\n""" ) print(f"""Height of Tree: {height(__lowerCamelCase )}""" , """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(__lowerCamelCase ) , """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1 , height(__lowerCamelCase ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(__lowerCamelCase , level=__lowerCamelCase ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
from __future__ import annotations import queue class __lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" snake_case_ = data snake_case_ = None snake_case_ = None def _lowerCAmelCase ( )->str: '''simple docstring''' print("\n********Press N to stop entering at any point of time********\n" ) snake_case_ = input("Enter the value of the root node: " ).strip().lower() snake_case_ = queue.Queue() snake_case_ = TreeNode(int(__lowerCamelCase ) ) q.put(__lowerCamelCase ) while not q.empty(): snake_case_ = q.get() snake_case_ = F'''Enter the left node of {node_found.data}: ''' snake_case_ = input(__lowerCamelCase ).strip().lower() or """n""" if check == "n": return tree_node snake_case_ = TreeNode(int(__lowerCamelCase ) ) snake_case_ = left_node q.put(__lowerCamelCase ) snake_case_ = F'''Enter the right node of {node_found.data}: ''' snake_case_ = input(__lowerCamelCase ).strip().lower() or """n""" if check == "n": return tree_node snake_case_ = TreeNode(int(__lowerCamelCase ) ) snake_case_ = right_node q.put(__lowerCamelCase ) raise def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->Any: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return print(node.data , end="," ) pre_order(node.left ) pre_order(node.right ) def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->Optional[int]: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return in_order(node.left ) print(node.data , end="," ) in_order(node.right ) def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->int: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end="," ) def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->int: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case_ = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): snake_case_ = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->int: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case_ = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): snake_case_ = [] while not q.empty(): snake_case_ = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__lowerCamelCase ) def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->Any: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case_ = [] snake_case_ = node while n or stack: while n: # start from root node, find its left child print(n.data , end="," ) stack.append(__lowerCamelCase ) snake_case_ = n.left # end of while means current node doesn't have left child snake_case_ = stack.pop() # start to traverse its right child snake_case_ = n.right def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->Tuple: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case_ = [] snake_case_ = node while n or stack: while n: stack.append(__lowerCamelCase ) snake_case_ = n.left snake_case_ = stack.pop() print(n.data , end="," ) snake_case_ = n.right def _lowerCAmelCase ( lowerCAmelCase_ :TreeNode )->Optional[Any]: '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case_ = [], [] snake_case_ = node stacka.append(__lowerCamelCase ) while stacka: # to find the reversed order of post order, store it in stack2 snake_case_ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__lowerCamelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end="," ) def _lowerCAmelCase ( lowerCAmelCase_ :str = "" , lowerCAmelCase_ :str=50 , lowerCAmelCase_ :Optional[Any]="*" )->Tuple: '''simple docstring''' if not s: return "\n" + width * char snake_case_ = divmod(width - len(__lowerCamelCase ) - 2 , 2 ) return F'''{left * char} {s} {(left + extra) * char}''' if __name__ == "__main__": import doctest doctest.testmod() print(prompt('''Binary Tree Traversals''')) SCREAMING_SNAKE_CASE :TreeNode = build_tree() print(prompt('''Pre Order Traversal''')) pre_order(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal''')) in_order(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal''')) post_order(node) print(prompt() + '''\n''') print(prompt('''Level Order Traversal''')) level_order(node) print(prompt() + '''\n''') print(prompt('''Actual Level Order Traversal''')) level_order_actual(node) print('''*''' * 50 + '''\n''') print(prompt('''Pre Order Traversal - Iteration Version''')) pre_order_iter(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal - Iteration Version''')) in_order_iter(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal - Iteration Version''')) post_order_iter(node) print(prompt())
283
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[int] = GPTSanJapaneseTokenizer a : Optional[Any] = False a : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off __UpperCAmelCase : Tuple = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __UpperCAmelCase : Dict = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __UpperCAmelCase : Dict = {"""unk_token""": """<unk>"""} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(__lowercase ) ) def UpperCAmelCase ( self : Tuple , **__lowercase : int ) -> Any: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def UpperCAmelCase ( self : str , __lowercase : Union[str, Any] ) -> Any: __UpperCAmelCase : Any = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __UpperCAmelCase : int = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase ( self : List[Any] , __lowercase : Optional[int] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase : int = self.get_input_output_texts(__lowercase ) __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : Dict = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) return text, ids def UpperCAmelCase ( self : int ) -> Optional[Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Dict ) -> Tuple: pass # TODO add if relevant def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : List[str] = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。 こんばんは、㔺界。""" __UpperCAmelCase : Dict = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids without special tokens __UpperCAmelCase : List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids with special tokens __UpperCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __UpperCAmelCase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : int = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : Tuple = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __UpperCAmelCase : int = """こんにちは、、、、世界。こんばんは、、、、世界。""" __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase ) __UpperCAmelCase : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : int ) -> Optional[int]: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : List[Any] = """こんにちは、世界。""" __UpperCAmelCase : Optional[int] = """こんばんは、㔺界。😀""" __UpperCAmelCase : List[Any] = """こんにちは、世界。こんばんは、世界。😀""" __UpperCAmelCase : List[str] = tokenizer.encode(prefix_text + input_text ) __UpperCAmelCase : List[Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __UpperCAmelCase : Any = tokenizer.encode(__lowercase , prefix_text=__lowercase ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowercase ) __UpperCAmelCase : Any = tokenizer.decode(__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。""" __UpperCAmelCase : List[Any] = """こんばんは、㔺界。😀""" __UpperCAmelCase : Union[str, Any] = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : int = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : List[Any] = [1] + [0] * (len_prefix + len_text + 1) __UpperCAmelCase : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] __UpperCAmelCase : List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCAmelCase : Union[str, Any] = tokenizer(prefix_text + input_text ).token_type_ids __UpperCAmelCase : Optional[Any] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __UpperCAmelCase : Tuple = tokenizer(__lowercase , prefix_text=__lowercase ).token_type_ids self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : List[str] ) -> int: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""あンいワ""" ) __UpperCAmelCase : Tuple = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertNotEqual(__lowercase , __lowercase ) self.assertNotEqual(__lowercase , __lowercase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self : List[Any] ) -> List[str]: __UpperCAmelCase : Any = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : List[Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __UpperCAmelCase : int = tokenizer(__lowercase , padding=__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.batch_encode_plus(__lowercase , padding=__lowercase ) # fmt: off __UpperCAmelCase : Optional[int] = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCAmelCase : Tuple = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCAmelCase : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __lowercase ) self.assertListEqual(x_token.token_type_ids , __lowercase ) self.assertListEqual(x_token.attention_mask , __lowercase ) self.assertListEqual(x_token_a.input_ids , __lowercase ) self.assertListEqual(x_token_a.token_type_ids , __lowercase ) self.assertListEqual(x_token_a.attention_mask , __lowercase ) def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase ( self : Any ) -> int: # tokenizer has no padding token pass
63
0
from typing import Any def __UpperCamelCase ( lowercase__ : list , lowercase__ : list , lowercase__ : dict , lowercase__ : dict , lowercase__ : dict , ) -> Union[str, Any]: '''simple docstring''' _validation( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) # Creates data structures and fill initial step lowerCAmelCase_ : dict = {} lowerCAmelCase_ : dict = {} for state in states_space: lowerCAmelCase_ : List[str] = observations_space[0] lowerCAmelCase_ : Union[str, Any] = ( initial_probabilities[state] * emission_probabilities[state][observation] ) lowerCAmelCase_ : Optional[int] = None # Fills the data structure with the probabilities of # different transitions and pointers to previous states for o in range(1 , len(__lowerCamelCase ) ): lowerCAmelCase_ : List[Any] = observations_space[o] lowerCAmelCase_ : Any = observations_space[o - 1] for state in states_space: # Calculates the argmax for probability function lowerCAmelCase_ : Optional[int] = """""" lowerCAmelCase_ : Tuple = -1 for k_state in states_space: lowerCAmelCase_ : str = ( probabilities[(k_state, prior_observation)] * transition_probabilities[k_state][state] * emission_probabilities[state][observation] ) if probability > max_probability: lowerCAmelCase_ : int = probability lowerCAmelCase_ : int = k_state # Update probabilities and pointers dicts lowerCAmelCase_ : Optional[Any] = ( probabilities[(arg_max, prior_observation)] * transition_probabilities[arg_max][state] * emission_probabilities[state][observation] ) lowerCAmelCase_ : Tuple = arg_max # The final observation lowerCAmelCase_ : int = observations_space[len(__lowerCamelCase ) - 1] # argmax for given final observation lowerCAmelCase_ : Tuple = """""" lowerCAmelCase_ : List[str] = -1 for k_state in states_space: lowerCAmelCase_ : Dict = probabilities[(k_state, final_observation)] if probability > max_probability: lowerCAmelCase_ : Dict = probability lowerCAmelCase_ : Tuple = k_state lowerCAmelCase_ : Tuple = arg_max # Process pointers backwards lowerCAmelCase_ : Tuple = last_state lowerCAmelCase_ : Dict = [] for o in range(len(__lowerCamelCase ) - 1 , -1 , -1 ): result.append(__lowerCamelCase ) lowerCAmelCase_ : str = pointers[previous, observations_space[o]] result.reverse() return result def __UpperCamelCase ( lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , ) -> Optional[Any]: '''simple docstring''' _validate_not_empty( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) _validate_lists(__lowerCamelCase , __lowerCamelCase ) _validate_dicts( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , ) -> int: '''simple docstring''' if not all( [ observations_space, states_space, initial_probabilities, transition_probabilities, emission_probabilities, ] ): raise ValueError("""There's an empty parameter""" ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : Any ) -> Optional[int]: '''simple docstring''' _validate_list(__lowerCamelCase , """observations_space""" ) _validate_list(__lowerCamelCase , """states_space""" ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : str ) -> Any: '''simple docstring''' if not isinstance(_object , __lowerCamelCase ): lowerCAmelCase_ : List[str] = f'{var_name} must be a list' raise ValueError(__lowerCamelCase ) else: for x in _object: if not isinstance(__lowerCamelCase , __lowerCamelCase ): lowerCAmelCase_ : Dict = f'{var_name} must be a list of strings' raise ValueError(__lowerCamelCase ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : Any , lowercase__ : Any , ) -> Optional[int]: '''simple docstring''' _validate_dict(__lowerCamelCase , """initial_probabilities""" , __lowerCamelCase ) _validate_nested_dict(__lowerCamelCase , """transition_probabilities""" ) _validate_nested_dict(__lowerCamelCase , """emission_probabilities""" ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : str ) -> List[str]: '''simple docstring''' _validate_dict(_object , __lowerCamelCase , __lowerCamelCase ) for x in _object.values(): _validate_dict(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def __UpperCamelCase ( lowercase__ : Any , lowercase__ : str , lowercase__ : type , lowercase__ : bool = False ) -> List[Any]: '''simple docstring''' if not isinstance(_object , __lowerCamelCase ): lowerCAmelCase_ : Any = f'{var_name} must be a dict' raise ValueError(__lowerCamelCase ) if not all(isinstance(__lowerCamelCase , __lowerCamelCase ) for x in _object ): lowerCAmelCase_ : Union[str, Any] = f'{var_name} all keys must be strings' raise ValueError(__lowerCamelCase ) if not all(isinstance(__lowerCamelCase , __lowerCamelCase ) for x in _object.values() ): lowerCAmelCase_ : int = """nested dictionary """ if nested else """""" lowerCAmelCase_ : Optional[Any] = f'{var_name} {nested_text}all values must be {value_type.__name__}' raise ValueError(__lowerCamelCase ) if __name__ == "__main__": from doctest import testmod testmod()
600
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a : Dict = logging.get_logger(__name__) @dataclass class a ( lowercase__ ): """simple docstring""" a : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : List[Any] , **__lowercase : Dict ) -> Tuple: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase : List[Any] = deprecated_arg[3:] setattr(self , __lowercase , not kwargs.pop(__lowercase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __UpperCAmelCase : str = kwargs.pop("""torchscript""" , self.torchscript ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) __UpperCAmelCase : Optional[Any] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**__lowercase ) a : bool = field(default=lowercase__ , metadata={'help': 'Trace the models using torchscript'} ) a : bool = field(default=lowercase__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) a : str = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def UpperCAmelCase ( self : Any ) -> Tuple["torch.device", int]: requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: __UpperCAmelCase : str = torch.device("""cpu""" ) __UpperCAmelCase : int = 0 elif is_torch_tpu_available(): __UpperCAmelCase : Tuple = xm.xla_device() __UpperCAmelCase : int = 0 else: __UpperCAmelCase : Dict = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) __UpperCAmelCase : Optional[int] = torch.cuda.device_count() return device, n_gpu @property def UpperCAmelCase ( self : Optional[Any] ) -> str: return is_torch_tpu_available() and self.tpu @property def UpperCAmelCase ( self : List[str] ) -> int: requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCAmelCase ( self : int ) -> "torch.device": requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def UpperCAmelCase ( self : int ) -> List[Any]: requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return self.n_gpu > 0
63
0
from torch import nn def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'''Unsupported activation function: {act_fn}''' )
6
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __UpperCAmelCase : str = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Any ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __UpperCAmelCase : Any = features.copy() if features else default_expected_features __UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=__lowerCamelCase , cache_dir=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: __UpperCAmelCase : Dict = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): __UpperCAmelCase : Optional[int] = tmp_path / """cache""" __UpperCAmelCase : str = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() __UpperCAmelCase : Optional[int] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Dict = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): __UpperCAmelCase : int = tmp_path / """cache""" __UpperCAmelCase : int = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Any = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Optional[int] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() with pytest.raises(__lowerCamelCase ): SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
63
0
from __future__ import annotations def lowerCAmelCase__ ( _a : int , _a : int ): if partitions <= 0: raise ValueError("partitions must be a positive number!" ) if partitions > number_of_bytes: raise ValueError("partitions can not > number_of_bytes!" ) snake_case_ : int = number_of_bytes // partitions snake_case_ : str = [] for i in range(__lowerCamelCase ): snake_case_ : Optional[Any] = i * bytes_per_partition + 1 snake_case_ : Any = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F'''{start_bytes}-{end_bytes}''' ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
568
from __future__ import annotations a : Optional[Any] = [True] * 1_000_001 a : Union[str, Any] = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): a : Optional[Any] = False i += 1 def lowerCamelCase__ ( __lowerCamelCase : int ): return seive[n] def lowerCamelCase__ ( __lowerCamelCase : int ): return any(digit in """02468""" for digit in str(__lowerCamelCase ) ) def lowerCamelCase__ ( __lowerCamelCase : int = 1000000 ): __UpperCAmelCase : Optional[Any] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__lowerCamelCase ) and not contains_an_even_digit(__lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) __UpperCAmelCase : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__lowerCamelCase ) )] if all(is_prime(__lowerCamelCase ) for i in list_nums ): result.append(__lowerCamelCase ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"""{len(find_circular_primes()) = }""")
63
0
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class SCREAMING_SNAKE_CASE_ ( lowercase__ , unittest.TestCase ): '''simple docstring''' lowercase : int = KandinskyVaaInpaintPipeline lowercase : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] lowercase : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] lowercase : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] lowercase : List[Any] = False @property def SCREAMING_SNAKE_CASE_ ( self : int ) -> Dict: return 32 @property def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> Optional[Any]: return 32 @property def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def SCREAMING_SNAKE_CASE_ ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> List[str]: return 1_00 @property def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> Any: torch.manual_seed(0 ) A : Optional[Any] ={ """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } A : int =UNetaDConditionModel(**__lowercase ) return model @property def SCREAMING_SNAKE_CASE_ ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> List[str]: torch.manual_seed(0 ) A : List[Any] =VQModel(**self.dummy_movq_kwargs ) return model def SCREAMING_SNAKE_CASE_ ( self : Any ) -> List[Any]: A : List[str] =self.dummy_unet A : List[str] =self.dummy_movq A : Optional[Any] =DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=__lowercase , ) A : str ={ """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def SCREAMING_SNAKE_CASE_ ( self : str , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str]=0 ) -> Optional[Any]: A : str =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) A : Optional[int] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image A : Tuple =floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) A : Union[str, Any] =image.cpu().permute(0 , 2 , 3 , 1 )[0] A : Union[str, Any] =Image.fromarray(np.uinta(__lowercase ) ).convert('RGB' ).resize((2_56, 2_56) ) # create mask A : Union[str, Any] =np.ones((64, 64) , dtype=np.floataa ) A : List[str] =0 if str(__lowercase ).startswith('mps' ): A : List[str] =torch.manual_seed(__lowercase ) else: A : Optional[int] =torch.Generator(device=__lowercase ).manual_seed(__lowercase ) A : Optional[Any] ={ """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def SCREAMING_SNAKE_CASE_ ( self : List[str] ) -> Union[str, Any]: A : Optional[Any] ="""cpu""" A : Dict =self.get_dummy_components() A : str =self.pipeline_class(**__lowercase ) A : Tuple =pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) A : List[Any] =pipe(**self.get_dummy_inputs(__lowercase ) ) A : Tuple =output.images A : Optional[int] =pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] A : Union[str, Any] =image[0, -3:, -3:, -1] A : str =image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}' ) assert image.shape == (1, 64, 64, 3) A : Optional[Any] =np.array( [0.5_0_7_7_5_9_0_3, 0.4_9_5_2_7_1_9_5, 0.4_8_8_2_4_5_4_3, 0.5_0_1_9_2_2_3_7, 0.4_8_6_4_4_9_0_6, 0.4_9_3_7_3_8_1_4, 0.4_7_8_0_5_9_8, 0.4_7_2_3_4_8_2_7, 0.4_8_3_2_7_8_4_8] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def SCREAMING_SNAKE_CASE_ ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Union[str, Any]: A : Any =load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy' ) A : List[str] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) A : List[Any] =np.ones((7_68, 7_68) , dtype=np.floataa ) A : Optional[Any] =0 A : Tuple ="""a hat""" A : str =KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) A : Any =KandinskyVaaInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder-inpaint' , torch_dtype=torch.floataa ) A : int =pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) A : int =torch.Generator(device='cpu' ).manual_seed(0 ) A : Optional[int] =pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() A : Optional[int] =pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) A : List[Any] =output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
305
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() a : Dict = logging.get_logger(__name__) a : Tuple = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowerCamelCase__ ( __lowerCamelCase : Dict ): if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase : Union[str, Any] = k.replace(__lowerCamelCase , __lowerCamelCase ) if k.startswith("""encoder""" ): __UpperCAmelCase : List[str] = k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase : Optional[Any] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : Union[str, Any] = k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase : Optional[int] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : List[Any] = k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase : Any = k.replace("""norm3""" , """final_layer_norm""" ) return k def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Optional[Any] = [ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase : Dict = sd.pop(__lowerCamelCase ) __UpperCAmelCase : List[str] = k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase : List[str] = v a : Optional[int] = ["START"] @torch.no_grad() def lowerCamelCase__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : str ): __UpperCAmelCase : str = torch.load(__lowerCamelCase , map_location="""cpu""" ) __UpperCAmelCase : Tuple = model["""model"""] __UpperCAmelCase : int = BlenderbotConfig.from_json_file(__lowerCamelCase ) __UpperCAmelCase : List[str] = BlenderbotForConditionalGeneration(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = m.model.state_dict().keys() __UpperCAmelCase : Any = [] __UpperCAmelCase : Any = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase : int = rename_state_dict_key(__lowerCamelCase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase : str = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__lowerCamelCase ) m.model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) m.half() m.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) a : Any = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
63
0
def _SCREAMING_SNAKE_CASE ( __lowercase : int ) -> Optional[Any]: """simple docstring""" __A = [1] __A = 0, 0, 0 __A = ugly_nums[ia] * 2 __A = ugly_nums[ia] * 3 __A = ugly_nums[ia] * 5 for _ in range(1 , __lowerCamelCase ): __A = min(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ugly_nums.append(__lowerCamelCase ) if next_num == next_a: ia += 1 __A = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 __A = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 __A = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f"""{ugly_numbers(200) = }""")
637
def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ): __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : List[str] = len(__lowerCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Union[str, Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None __UpperCAmelCase : str = sorted_collection[point] if current_item == item: return point else: if point < left: __UpperCAmelCase : Optional[Any] = left __UpperCAmelCase : Tuple = point elif point > right: __UpperCAmelCase : Optional[Any] = right __UpperCAmelCase : Dict = point else: if item < current_item: __UpperCAmelCase : Union[str, Any] = point - 1 else: __UpperCAmelCase : str = point + 1 return None def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) elif point > right: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , point - 1 ) else: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , point + 1 , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : int ): if collection != sorted(__lowerCamelCase ): raise ValueError("""Collection must be ascending sorted""" ) return True if __name__ == "__main__": import sys a : Optional[Any] = 0 if debug == 1: a : Optional[Any] = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") a : Tuple = 67 a : List[Any] = interpolation_search(collection, target) if result is not None: print(f"""{target} found at positions: {result}""") else: print("Not found")
63
0
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class a__ : def __init__( self :Any , _lowerCamelCase :Dict , _lowerCamelCase :int = 13 , _lowerCamelCase :int = 64 , _lowerCamelCase :int = 2 , _lowerCamelCase :int = 3 , _lowerCamelCase :int = 3 , _lowerCamelCase :bool = True , _lowerCamelCase :bool = True , _lowerCamelCase :int = 128 , _lowerCamelCase :List[str]=[16, 32, 64, 128] , _lowerCamelCase :int = 7 , _lowerCamelCase :int = 4 , _lowerCamelCase :int = 37 , _lowerCamelCase :str = "gelu" , _lowerCamelCase :float = 0.1 , _lowerCamelCase :float = 0.1 , _lowerCamelCase :int = 10 , _lowerCamelCase :float = 0.02 , _lowerCamelCase :int = 2 , _lowerCamelCase :int = 1 , _lowerCamelCase :int = 128 , _lowerCamelCase :List[int] = [2, 2, 2, 2] , _lowerCamelCase :int = 2 , _lowerCamelCase :int = 2 , ): '''simple docstring''' UpperCamelCase_ : Optional[int] =parent UpperCamelCase_ : Tuple =batch_size UpperCamelCase_ : int =image_size UpperCamelCase_ : Optional[int] =patch_size UpperCamelCase_ : Any =num_channels UpperCamelCase_ : List[str] =is_training UpperCamelCase_ : Optional[Any] =use_labels UpperCamelCase_ : List[Any] =hidden_size UpperCamelCase_ : Optional[int] =num_hidden_layers UpperCamelCase_ : Union[str, Any] =num_attention_heads UpperCamelCase_ : Union[str, Any] =intermediate_size UpperCamelCase_ : List[Any] =hidden_act UpperCamelCase_ : Any =hidden_dropout_prob UpperCamelCase_ : Dict =attention_probs_dropout_prob UpperCamelCase_ : Union[str, Any] =type_sequence_label_size UpperCamelCase_ : Dict =initializer_range UpperCamelCase_ : str =encoder_stride UpperCamelCase_ : List[Any] =num_attention_outputs UpperCamelCase_ : str =embed_dim UpperCamelCase_ : Tuple =embed_dim + 1 UpperCamelCase_ : Tuple =resolution UpperCamelCase_ : Union[str, Any] =depths UpperCamelCase_ : List[Any] =hidden_sizes UpperCamelCase_ : Optional[Any] =dim UpperCamelCase_ : Union[str, Any] =mlp_expansion_ratio def lowerCamelCase_ ( self :str ): '''simple docstring''' UpperCamelCase_ : Optional[Any] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase_ : int =None if self.use_labels: UpperCamelCase_ : Tuple =ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase_ : Any =self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self :List[Any] ): '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCamelCase_ ( self :Optional[int] , _lowerCamelCase :int , _lowerCamelCase :List[str] , _lowerCamelCase :Optional[Any] ): '''simple docstring''' UpperCamelCase_ : Tuple =TFEfficientFormerModel(config=__lowercase ) UpperCamelCase_ : int =model(__lowercase , training=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self :Union[str, Any] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :Dict , _lowerCamelCase :Any ): '''simple docstring''' UpperCamelCase_ : Tuple =self.type_sequence_label_size UpperCamelCase_ : Union[str, Any] =TFEfficientFormerForImageClassification(__lowercase ) UpperCamelCase_ : List[str] =model(__lowercase , labels=__lowercase , training=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase_ : Dict =1 UpperCamelCase_ : Optional[Any] =TFEfficientFormerForImageClassification(__lowercase ) UpperCamelCase_ : Dict =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase_ : str =model(__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase_ ( self :Optional[Any] ): '''simple docstring''' UpperCamelCase_ : Tuple =self.prepare_config_and_inputs() UpperCamelCase_ : Union[str, Any] =config_and_inputs UpperCamelCase_ : Union[str, Any] ={"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class a__ ( lowercase__ , lowercase__ , unittest.TestCase ): UpperCAmelCase__ = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) UpperCAmelCase__ = ( { 'feature-extraction': TFEfficientFormerModel, 'image-classification': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def lowerCamelCase_ ( self :Tuple ): '''simple docstring''' UpperCamelCase_ : int =TFEfficientFormerModelTester(self ) UpperCamelCase_ : Optional[int] =ConfigTester( self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37 ) def lowerCamelCase_ ( self :Optional[int] ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='EfficientFormer does not use inputs_embeds' ) def lowerCamelCase_ ( self :Any ): '''simple docstring''' pass @unittest.skip(reason='EfficientFormer does not support input and output embeddings' ) def lowerCamelCase_ ( self :Dict ): '''simple docstring''' pass def lowerCamelCase_ ( self :Tuple ): '''simple docstring''' UpperCamelCase_ : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ : List[Any] =model_class(__lowercase ) UpperCamelCase_ : Optional[int] =inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase_ : Optional[int] =[*signature.parameters.keys()] UpperCamelCase_ : Optional[int] =["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowercase ) def lowerCamelCase_ ( self :List[Any] ): '''simple docstring''' def check_hidden_states_output(_lowerCamelCase :Optional[int] , _lowerCamelCase :Union[str, Any] , _lowerCamelCase :Optional[int] ): UpperCamelCase_ : Optional[int] =model_class(__lowercase ) UpperCamelCase_ : Any =model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) UpperCamelCase_ : Union[str, Any] =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase_ : int =getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__lowercase ) , __lowercase ) if hasattr(self.model_tester , 'encoder_seq_length' ): UpperCamelCase_ : List[str] =self.model_tester.encoder_seq_length if hasattr(self.model_tester , 'chunk_length' ) and self.model_tester.chunk_length > 1: UpperCamelCase_ : Any =seq_length * self.model_tester.chunk_length else: UpperCamelCase_ : str =self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: UpperCamelCase_ : str =outputs.decoder_hidden_states self.asseretIsInstance(__lowercase , (list, tuple) ) self.assertEqual(len(__lowercase ) , __lowercase ) UpperCamelCase_ : Dict =getattr(self.model_tester , 'seq_length' , __lowercase ) UpperCamelCase_ : str =getattr(self.model_tester , 'decoder_seq_length' , __lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) UpperCamelCase_ : List[Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ : Optional[Any] =True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase_ : Optional[int] =True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) def lowerCamelCase_ ( self :str , _lowerCamelCase :str , _lowerCamelCase :Optional[int] , _lowerCamelCase :Union[str, Any]=False ): '''simple docstring''' UpperCamelCase_ : Optional[Any] =super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCamelCase_ ( self :str ): '''simple docstring''' UpperCamelCase_ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) @unittest.skip(reason='EfficientFormer does not implement masked image modeling yet' ) def lowerCamelCase_ ( self :str ): '''simple docstring''' UpperCamelCase_ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowercase ) def lowerCamelCase_ ( self :Any ): '''simple docstring''' UpperCamelCase_ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) @slow def lowerCamelCase_ ( self :int ): '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase_ : Union[str, Any] =TFEfficientFormerModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def lowerCamelCase_ ( self :Optional[int] ): '''simple docstring''' UpperCamelCase_ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ : Tuple =True UpperCamelCase_ : Optional[Any] =getattr(self.model_tester , 'seq_length' , __lowercase ) UpperCamelCase_ : Optional[Any] =getattr(self.model_tester , 'encoder_seq_length' , __lowercase ) UpperCamelCase_ : Any =getattr(self.model_tester , 'key_length' , __lowercase ) UpperCamelCase_ : Optional[Any] =getattr(self.model_tester , 'chunk_length' , __lowercase ) if chunk_length is not None and hasattr(self.model_tester , 'num_hashes' ): UpperCamelCase_ : Dict =encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: UpperCamelCase_ : List[str] =True UpperCamelCase_ : List[Any] =False UpperCamelCase_ : Dict =True UpperCamelCase_ : Dict =model_class(__lowercase ) UpperCamelCase_ : Optional[Any] =model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) UpperCamelCase_ : int =outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCamelCase_ : Optional[int] =True UpperCamelCase_ : Any =model_class(__lowercase ) UpperCamelCase_ : str =model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) UpperCamelCase_ : Tuple =outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCamelCase_ ( self :List[Any] ): '''simple docstring''' UpperCamelCase_ : str =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model UpperCamelCase_ : int =model_class(__lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes UpperCamelCase_ : Tuple ={ key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=__lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } UpperCamelCase_ : str =model(__lowercase ) self.assertTrue(outputs_dict is not None ) def A_ ( ): UpperCamelCase_ : str =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class a__ ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self :Any ): '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained('snap-research/efficientformer-l1-300' ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self :Optional[int] ): '''simple docstring''' UpperCamelCase_ : Any =TFEfficientFormerForImageClassification.from_pretrained('snap-research/efficientformer-l1-300' ) UpperCamelCase_ : Optional[Any] =self.default_image_processor UpperCamelCase_ : Optional[Any] =prepare_img() UpperCamelCase_ : Optional[Any] =image_processor(images=__lowercase , return_tensors='tf' ) # forward pass UpperCamelCase_ : Optional[Any] =model(**__lowercase , training=__lowercase ) # verify the logits UpperCamelCase_ : Any =tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowercase ) UpperCamelCase_ : Optional[int] =tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) ) @slow def lowerCamelCase_ ( self :Tuple ): '''simple docstring''' UpperCamelCase_ : Tuple =TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( 'snap-research/efficientformer-l1-300' ) UpperCamelCase_ : Dict =self.default_image_processor UpperCamelCase_ : int =prepare_img() UpperCamelCase_ : Dict =image_processor(images=__lowercase , return_tensors='tf' ) # forward pass UpperCamelCase_ : Any =model(**__lowercase , training=__lowercase ) # verify the logits UpperCamelCase_ : Any =tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowercase ) UpperCamelCase_ : List[str] =tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) )
357
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
63
0
__snake_case : int = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" __snake_case : Optional[Any] = [{"type": "code", "content": INSTALL_CONTENT}] __snake_case : Optional[int] = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
540
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class a ( lowercase__ , lowercase__ ): """simple docstring""" a : Dict = 1 @register_to_config def __init__( self : int , __lowercase : int = 1000 , __lowercase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: # set `betas`, `alphas`, `timesteps` self.set_timesteps(__lowercase ) # standard deviation of the initial noise distribution __UpperCAmelCase : List[Any] = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __UpperCAmelCase : List[Any] = 4 # running values __UpperCAmelCase : str = [] def UpperCAmelCase ( self : Union[str, Any] , __lowercase : int , __lowercase : Union[str, torch.device] = None ) -> int: __UpperCAmelCase : int = num_inference_steps __UpperCAmelCase : Union[str, Any] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __UpperCAmelCase : Union[str, Any] = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __UpperCAmelCase : Dict = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __UpperCAmelCase : Dict = torch.sin(steps * math.pi / 2 ) ** 2 __UpperCAmelCase : List[Any] = (1.0 - self.betas**2) ** 0.5 __UpperCAmelCase : Tuple = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __UpperCAmelCase : Dict = timesteps.to(__lowercase ) __UpperCAmelCase : Optional[Any] = [] def UpperCAmelCase ( self : Optional[int] , __lowercase : torch.FloatTensor , __lowercase : int , __lowercase : torch.FloatTensor , __lowercase : bool = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __UpperCAmelCase : List[str] = (self.timesteps == timestep).nonzero().item() __UpperCAmelCase : Optional[Any] = timestep_index + 1 __UpperCAmelCase : List[str] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowercase ) if len(self.ets ) == 1: __UpperCAmelCase : Tuple = self.ets[-1] elif len(self.ets ) == 2: __UpperCAmelCase : Union[str, Any] = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __UpperCAmelCase : Union[str, Any] = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __UpperCAmelCase : List[Any] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __UpperCAmelCase : Union[str, Any] = self._get_prev_sample(__lowercase , __lowercase , __lowercase , __lowercase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowercase ) def UpperCAmelCase ( self : Optional[Any] , __lowercase : torch.FloatTensor , *__lowercase : Optional[Any] , **__lowercase : Any ) -> torch.FloatTensor: return sample def UpperCAmelCase ( self : Tuple , __lowercase : Tuple , __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Dict ) -> str: __UpperCAmelCase : int = self.alphas[timestep_index] __UpperCAmelCase : Tuple = self.betas[timestep_index] __UpperCAmelCase : Any = self.alphas[prev_timestep_index] __UpperCAmelCase : List[str] = self.betas[prev_timestep_index] __UpperCAmelCase : List[str] = (sample - sigma * ets) / max(__lowercase , 1e-8 ) __UpperCAmelCase : List[Any] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Tuple ) -> str: return self.config.num_train_timesteps
63
0
'''simple docstring''' def __snake_case ( lowerCamelCase_ : float , lowerCamelCase_ : int ): '''simple docstring''' if digit_amount > 0: return round(number - int(__lowerCamelCase ) , __lowerCamelCase ) return number - int(__lowerCamelCase ) if __name__ == "__main__": print(decimal_isolate(1.5_3, 0)) print(decimal_isolate(3_5.3_4_5, 1)) print(decimal_isolate(3_5.3_4_5, 2)) print(decimal_isolate(3_5.3_4_5, 3)) print(decimal_isolate(-1_4.7_8_9, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-1_4.1_2_3, 1)) print(decimal_isolate(-1_4.1_2_3, 2)) print(decimal_isolate(-1_4.1_2_3, 3))
664
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCamelCase__ ( ): __UpperCAmelCase : Union[str, Any] = ArgumentParser("""Transformers CLI tool""" , usage="""transformers-cli <command> [<args>]""" ) __UpperCAmelCase : Any = parser.add_subparsers(help="""transformers-cli command helpers""" ) # Register commands ConvertCommand.register_subcommand(__lowerCamelCase ) DownloadCommand.register_subcommand(__lowerCamelCase ) EnvironmentCommand.register_subcommand(__lowerCamelCase ) RunCommand.register_subcommand(__lowerCamelCase ) ServeCommand.register_subcommand(__lowerCamelCase ) UserCommands.register_subcommand(__lowerCamelCase ) AddNewModelCommand.register_subcommand(__lowerCamelCase ) AddNewModelLikeCommand.register_subcommand(__lowerCamelCase ) LfsCommands.register_subcommand(__lowerCamelCase ) PTtoTFCommand.register_subcommand(__lowerCamelCase ) # Let's go __UpperCAmelCase : Optional[Any] = parser.parse_args() if not hasattr(__lowerCamelCase , """func""" ): parser.print_help() exit(1 ) # Run __UpperCAmelCase : Tuple = args.func(__lowerCamelCase ) service.run() if __name__ == "__main__": main()
63
0
import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all BART models at https://huggingface.co/models?filter=bart _UpperCamelCase = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, } _UpperCamelCase = { "facebook/bart-base": 1024, "facebook/bart-large": 1024, "facebook/bart-large-mnli": 1024, "facebook/bart-large-cnn": 1024, "facebook/bart-large-xsum": 1024, "yjernite/bart_eli5": 1024, } @lru_cache() def UpperCamelCase_( ) -> Union[str, Any]: UpperCAmelCase__ = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) UpperCAmelCase__ = bs[:] UpperCAmelCase__ = 0 for b in range(2**8 ): if b not in bs: bs.append(__lowerCamelCase ) cs.append(2**8 + n ) n += 1 UpperCAmelCase__ = [chr(__lowerCamelCase ) for n in cs] return dict(zip(__lowerCamelCase , __lowerCamelCase ) ) def UpperCamelCase_( snake_case__: int ) -> List[Any]: UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs class lowercase ( lowercase__ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = ['input_ids', 'attention_mask'] def __init__(self , __a , __a , __a="replace" , __a="<s>" , __a="</s>" , __a="</s>" , __a="<s>" , __a="<unk>" , __a="<pad>" , __a="<mask>" , __a=False , **__a , ) -> Any: """simple docstring""" UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else bos_token UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else eos_token UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else sep_token UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else cls_token UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else unk_token UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token super().__init__( errors=__lowercase , bos_token=__lowercase , eos_token=__lowercase , unk_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , add_prefix_space=__lowercase , **__lowercase , ) with open(__lowercase , encoding='utf-8' ) as vocab_handle: UpperCAmelCase__ = json.load(__lowercase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} UpperCAmelCase__ = errors # how to handle errors in decoding UpperCAmelCase__ = bytes_to_unicode() UpperCAmelCase__ = {v: k for k, v in self.byte_encoder.items()} with open(__lowercase , encoding='utf-8' ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split('\n' )[1:-1] UpperCAmelCase__ = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase__ = dict(zip(__lowercase , range(len(__lowercase ) ) ) ) UpperCAmelCase__ = {} UpperCAmelCase__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase__ = re.compile(r'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" return len(self.encoder ) def UpperCamelCase__ (self ) -> int: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def UpperCamelCase__ (self , __a ) -> Optional[int]: """simple docstring""" if token in self.cache: return self.cache[token] UpperCAmelCase__ = tuple(__lowercase ) UpperCAmelCase__ = get_pairs(__lowercase ) if not pairs: return token while True: UpperCAmelCase__ = min(__lowercase , key=lambda __a : self.bpe_ranks.get(__lowercase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__lowercase ): try: UpperCAmelCase__ = word.index(__lowercase , __lowercase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__lowercase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__lowercase ) UpperCAmelCase__ = new_word if len(__lowercase ) == 1: break else: UpperCAmelCase__ = get_pairs(__lowercase ) UpperCAmelCase__ = """ """.join(__lowercase ) UpperCAmelCase__ = word return word def UpperCamelCase__ (self , __a ) -> List[str]: """simple docstring""" UpperCAmelCase__ = [] for token in re.findall(self.pat , __lowercase ): UpperCAmelCase__ = """""".join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(__lowercase ).split(' ' ) ) return bpe_tokens def UpperCamelCase__ (self , __a ) -> str: """simple docstring""" return self.encoder.get(__lowercase , self.encoder.get(self.unk_token ) ) def UpperCamelCase__ (self , __a ) -> Optional[Any]: """simple docstring""" return self.decoder.get(__lowercase ) def UpperCamelCase__ (self , __a ) -> List[str]: """simple docstring""" UpperCAmelCase__ = """""".join(__lowercase ) UpperCAmelCase__ = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def UpperCamelCase__ (self , __a , __a = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__lowercase ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase__ = os.path.join( __lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join( __lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(__lowercase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__lowercase , ensure_ascii=__lowercase ) + '\n' ) UpperCAmelCase__ = 0 with open(__lowercase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __a : kv[1] ): if index != token_index: logger.warning( F"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." ' Please check that the tokenizer is not corrupted!' ) UpperCAmelCase__ = token_index writer.write(' '.join(__lowercase ) + '\n' ) index += 1 return vocab_file, merge_file def UpperCamelCase__ (self , __a , __a = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] UpperCAmelCase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase__ (self , __a , __a = None , __a = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase ) if token_ids_a is None: return [1] + ([0] * len(__lowercase )) + [1] return [1] + ([0] * len(__lowercase )) + [1, 1] + ([0] * len(__lowercase )) + [1] def UpperCamelCase__ (self , __a , __a = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ (self , __a , __a=False , **__a ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(__lowercase ) > 0 and not text[0].isspace()): UpperCAmelCase__ = """ """ + text return (text, kwargs)
146
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : List[str] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Dict = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
63
0
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" __lowercase = """""" for i in table: res += inp[i - 1] return res def _A ( A__ ): """simple docstring""" return data[1:] + data[0] def _A ( A__ , A__ ): """simple docstring""" __lowercase = """""" for i in range(len(__lowerCamelCase ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def _A ( A__ , A__ ): """simple docstring""" __lowercase = int('''0b''' + data[0] + data[-1] , 2 ) __lowercase = int('''0b''' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = message[:4] __lowercase = message[4:] __lowercase = apply_table(__lowerCamelCase , __lowerCamelCase ) __lowercase = xor(__lowerCamelCase , __lowerCamelCase ) __lowercase = apply_sbox(__lowerCamelCase , temp[:4] ) # noqa: E741 __lowercase = apply_sbox(__lowerCamelCase , temp[4:] ) __lowercase = """0""" * (2 - len(__lowerCamelCase )) + l # noqa: E741 __lowercase = """0""" * (2 - len(__lowerCamelCase )) + r __lowercase = apply_table(l + r , __lowerCamelCase ) __lowercase = xor(__lowerCamelCase , __lowerCamelCase ) return temp + right if __name__ == "__main__": lowerCAmelCase__ = input('''Enter 10 bit key: ''') lowerCAmelCase__ = input('''Enter 8 bit message: ''') lowerCAmelCase__ = [6, 3, 7, 4, 8, 5, 10, 9] lowerCAmelCase__ = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] lowerCAmelCase__ = [2, 4, 3, 1] lowerCAmelCase__ = [2, 6, 3, 1, 4, 8, 5, 7] lowerCAmelCase__ = [4, 1, 3, 5, 7, 2, 8, 6] lowerCAmelCase__ = [4, 1, 2, 3, 2, 3, 4, 1] lowerCAmelCase__ = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] lowerCAmelCase__ = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation lowerCAmelCase__ = apply_table(key, paa_table) lowerCAmelCase__ = temp[:5] lowerCAmelCase__ = temp[5:] lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = apply_table(left + right, pa_table) lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = apply_table(left + right, pa_table) # encryption lowerCAmelCase__ = apply_table(message, IP) lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = temp[4:] + temp[:4] lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = apply_table(temp, IP_inv) print('''Cipher text is:''', CT) # decryption lowerCAmelCase__ = apply_table(CT, IP) lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = temp[4:] + temp[:4] lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = apply_table(temp, IP_inv) print('''Plain text after decypting is:''', PT)
41
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys a : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. SCREAMING_SNAKE_CASE :Union[str, Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def _lowerCAmelCase ( lowerCAmelCase_ :Union[str, Any] )->Optional[int]: '''simple docstring''' from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(__lowerCamelCase ) def _lowerCAmelCase ( lowerCAmelCase_ :int )->int: '''simple docstring''' from diffusers.utils.testing_utils import pytest_terminal_summary_main snake_case_ = terminalreporter.config.getoption("--make-reports" ) if make_reports: pytest_terminal_summary_main(__lowerCamelCase , id=__lowerCamelCase )
283
def lowerCamelCase__ ( __lowerCamelCase : int ): if num <= 0: raise ValueError("""Input must be a positive integer""" ) __UpperCAmelCase : int = [True] * (num + 1) __UpperCAmelCase : Tuple = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCamelCase ): __UpperCAmelCase : str = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() a : Any = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
63
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
600
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Union[str, Any] = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class a ( lowercase__ ): """simple docstring""" a : Optional[int] = 'git_vision_model' def __init__( self : str , __lowercase : List[str]=768 , __lowercase : List[str]=3072 , __lowercase : List[Any]=12 , __lowercase : Dict=12 , __lowercase : int=3 , __lowercase : Any=224 , __lowercase : Optional[int]=16 , __lowercase : Dict="quick_gelu" , __lowercase : Any=1e-5 , __lowercase : str=0.0 , __lowercase : int=0.02 , **__lowercase : int , ) -> List[str]: super().__init__(**__lowercase ) __UpperCAmelCase : int = hidden_size __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Dict = num_attention_heads __UpperCAmelCase : int = num_channels __UpperCAmelCase : str = patch_size __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : int = initializer_range __UpperCAmelCase : Tuple = attention_dropout __UpperCAmelCase : Optional[int] = layer_norm_eps __UpperCAmelCase : List[Any] = hidden_act @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : Union[str, os.PathLike] , **__lowercase : List[str] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__lowercase ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("""model_type""" ) == "git": __UpperCAmelCase : str = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a ( lowercase__ ): """simple docstring""" a : List[str] = 'git' def __init__( self : Optional[int] , __lowercase : List[Any]=None , __lowercase : Tuple=30522 , __lowercase : str=768 , __lowercase : Optional[int]=6 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[int]=3072 , __lowercase : List[str]="gelu" , __lowercase : Tuple=0.1 , __lowercase : Optional[int]=0.1 , __lowercase : Optional[int]=1024 , __lowercase : Union[str, Any]=0.02 , __lowercase : Optional[Any]=1e-1_2 , __lowercase : List[Any]=0 , __lowercase : Dict="absolute" , __lowercase : Dict=True , __lowercase : Any=False , __lowercase : Optional[int]=101 , __lowercase : str=102 , __lowercase : Union[str, Any]=None , **__lowercase : Dict , ) -> Tuple: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: __UpperCAmelCase : Optional[int] = {} logger.info("""vision_config is None. initializing the GitVisionConfig with default values.""" ) __UpperCAmelCase : Tuple = GitVisionConfig(**__lowercase ) __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Union[str, Any] = hidden_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : List[str] = hidden_act __UpperCAmelCase : str = intermediate_size __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : int = max_position_embeddings __UpperCAmelCase : str = initializer_range __UpperCAmelCase : str = layer_norm_eps __UpperCAmelCase : Union[str, Any] = position_embedding_type __UpperCAmelCase : Dict = use_cache __UpperCAmelCase : int = tie_word_embeddings __UpperCAmelCase : Optional[int] = num_image_with_embedding __UpperCAmelCase : Optional[int] = bos_token_id __UpperCAmelCase : List[Any] = eos_token_id def UpperCAmelCase ( self : str ) -> int: __UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : List[str] = self.vision_config.to_dict() __UpperCAmelCase : Union[str, Any] = self.__class__.model_type return output
63
0
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _lowerCamelCase = "\\n@article{wang2019superglue,\n title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},\n author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R},\n journal={arXiv preprint arXiv:1905.00537},\n year={2019}\n}\n" _lowerCamelCase = "\\nSuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after\nGLUE with a new set of more difficult language understanding tasks, improved\nresources, and a new public leaderboard.\n" _lowerCamelCase = "\nCompute SuperGLUE evaluation metric associated to each SuperGLUE dataset.\nArgs:\n predictions: list of predictions to score. Depending on the SuperGlUE subset:\n - for 'record': list of question-answer dictionaries with the following keys:\n - 'idx': index of the question as specified by the dataset\n - 'prediction_text': the predicted answer text\n - for 'multirc': list of question-answer dictionaries with the following keys:\n - 'idx': index of the question-answer pair as specified by the dataset\n - 'prediction': the predicted answer label\n - otherwise: list of predicted labels\n references: list of reference labels. Depending on the SuperGLUE subset:\n - for 'record': list of question-answers dictionaries with the following keys:\n - 'idx': index of the question as specified by the dataset\n - 'answers': list of possible answers\n - otherwise: list of reference labels\nReturns: depending on the SuperGLUE subset:\n - for 'record':\n - 'exact_match': Exact match between answer and gold answer\n - 'f1': F1 score\n - for 'multirc':\n - 'exact_match': Exact match between answer and gold answer\n - 'f1_m': Per-question macro-F1 score\n - 'f1_a': Average F1 score over all answers\n - for 'axb':\n 'matthews_correlation': Matthew Correlation\n - for 'cb':\n - 'accuracy': Accuracy\n - 'f1': F1 score\n - for all others:\n - 'accuracy': Accuracy\nExamples:\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"]\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'cb')\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'record')\n >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}]\n >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 1.0, 'f1': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc')\n >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'axb')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any , UpperCamelCase__: Any ): return float((preds == labels).mean() ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] , UpperCamelCase__: Any , UpperCamelCase__: int="binary" ): SCREAMING_SNAKE_CASE__ = simple_accuracy(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = float(fa_score(y_true=__lowerCamelCase , y_pred=__lowerCamelCase , average=__lowerCamelCase ) ) return { "accuracy": acc, "f1": fa, } def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = {} for id_pred, label in zip(__lowerCamelCase , __lowerCamelCase ): SCREAMING_SNAKE_CASE__ = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' SCREAMING_SNAKE_CASE__ = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: SCREAMING_SNAKE_CASE__ = [(pred, label)] SCREAMING_SNAKE_CASE__ = [], [] for question, preds_labels in question_map.items(): SCREAMING_SNAKE_CASE__ = zip(*__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = fa_score(y_true=__lowerCamelCase , y_pred=__lowerCamelCase , average="""macro""" ) fas.append(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = int(sum(pred == label for pred, label in preds_labels ) == len(__lowerCamelCase ) ) ems.append(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = float(sum(__lowerCamelCase ) / len(__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__ = sum(__lowerCamelCase ) / len(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = float(fa_score(y_true=__lowerCamelCase , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase_ ( datasets.Metric ): def _snake_case ( self :Optional[Any] ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _snake_case ( self :Optional[Any] ) -> Optional[int]: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _snake_case ( self :Any , __A :Dict , __A :str ) -> int: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowercase , __lowercase )} elif self.config_name == "cb": return acc_and_fa(__lowercase , __lowercase , fa_avg="""macro""" ) elif self.config_name == "record": SCREAMING_SNAKE_CASE__ = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] SCREAMING_SNAKE_CASE__ = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowercase , __lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowercase , __lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowercase , __lowercase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
6
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = BarthezTokenizer a : Any = BarthezTokenizerFast a : Union[str, Any] = True a : Union[str, Any] = True def UpperCAmelCase ( self : Dict ) -> Any: super().setUp() __UpperCAmelCase : Optional[int] = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=__lowercase ) __UpperCAmelCase : str = tokenizer def UpperCAmelCase ( self : Optional[int] ) -> Tuple: __UpperCAmelCase : Dict = """<pad>""" __UpperCAmelCase : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> str: __UpperCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__lowercase ) , 101122 ) def UpperCAmelCase ( self : Any ) -> List[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __UpperCAmelCase : str = [0, 57, 3018, 70307, 91, 2] __UpperCAmelCase : List[Any] = self.tokenizer( __lowercase , max_length=len(__lowercase ) , padding=__lowercase , truncation=__lowercase , return_tensors="""pt""" ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __UpperCAmelCase : int = batch.input_ids.tolist()[0] self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> Tuple: if not self.test_rust_tokenizer: return __UpperCAmelCase : Union[str, Any] = self.get_tokenizer() __UpperCAmelCase : Optional[Any] = self.get_rust_tokenizer() __UpperCAmelCase : int = """I was born in 92000, and this is falsé.""" __UpperCAmelCase : Union[str, Any] = tokenizer.tokenize(__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = self.get_rust_tokenizer() __UpperCAmelCase : str = tokenizer.encode(__lowercase ) __UpperCAmelCase : Tuple = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: # fmt: off __UpperCAmelCase : str = {"""input_ids""": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __UpperCAmelCase : int = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=__lowercase , )
63
0
from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def lowerCAmelCase__ ( _a : str , _a : str , _a : Optional[str] = None ): if version.parse(hfh.__version__ ).release < version.parse("0.11.0" ).release: # old versions of hfh don't url-encode the file path snake_case_ : Tuple = quote(__lowerCamelCase ) return hfh.hf_hub_url(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" , revision=__lowerCamelCase )
568
from __future__ import annotations import math def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : bool , __lowerCamelCase : list[int] , __lowerCamelCase : float ): if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(__lowerCamelCase ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = [90, 23, 6, 33, 21, 65, 123, 34423] __UpperCAmelCase : str = math.log(len(__lowerCamelCase ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class SCREAMING_SNAKE_CASE_ ( lowercase__ ): '''simple docstring''' def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[Any]=13 , SCREAMING_SNAKE_CASE__ : List[str]=7 , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : Tuple=True , SCREAMING_SNAKE_CASE__ : List[Any]=False , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : List[str]=99 , SCREAMING_SNAKE_CASE__ : Any=32 , SCREAMING_SNAKE_CASE__ : Optional[Any]=5 , SCREAMING_SNAKE_CASE__ : Dict=4 , SCREAMING_SNAKE_CASE__ : Optional[Any]=37 , SCREAMING_SNAKE_CASE__ : int="gelu" , SCREAMING_SNAKE_CASE__ : Optional[int]=0.1 , SCREAMING_SNAKE_CASE__ : List[Any]=0.1 , SCREAMING_SNAKE_CASE__ : Tuple=5_12 , SCREAMING_SNAKE_CASE__ : str=16 , SCREAMING_SNAKE_CASE__ : Dict=2 , SCREAMING_SNAKE_CASE__ : Any=0.0_2 , SCREAMING_SNAKE_CASE__ : Optional[Any]=3 , SCREAMING_SNAKE_CASE__ : Any=4 , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , ) -> List[Any]: A : str =parent A : Optional[Any] =batch_size A : Tuple =seq_length A : Union[str, Any] =is_training A : Any =use_input_mask A : Optional[int] =use_token_type_ids A : Optional[int] =use_labels A : Optional[Any] =vocab_size A : Dict =hidden_size A : Optional[int] =num_hidden_layers A : int =num_attention_heads A : int =intermediate_size A : Optional[int] =hidden_act A : List[Any] =hidden_dropout_prob A : List[str] =attention_probs_dropout_prob A : List[str] =max_position_embeddings A : List[str] =type_vocab_size A : Tuple =type_sequence_label_size A : int =initializer_range A : Optional[int] =num_labels A : Any =num_choices A : Tuple =scope def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Union[str, Any]: A : List[Any] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A : str =None if self.use_input_mask: A : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) A : Optional[Any] =None A : str =None A : Optional[Any] =None if self.use_labels: A : Optional[int] =ids_tensor([self.batch_size] , self.type_sequence_label_size ) A : List[Any] =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A : Union[str, Any] =ids_tensor([self.batch_size] , self.num_choices ) A : Any =self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE_ ( self : Any ) -> int: return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE_ ( self : Any , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : str ) -> str: A : Any =DistilBertModel(config=__lowercase ) model.to(__lowercase ) model.eval() A : Any =model(__lowercase , __lowercase ) A : Union[str, Any] =model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Dict ) -> int: A : Union[str, Any] =DistilBertForMaskedLM(config=__lowercase ) model.to(__lowercase ) model.eval() A : Dict =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE_ ( self : str , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] ) -> Tuple: A : Tuple =DistilBertForQuestionAnswering(config=__lowercase ) model.to(__lowercase ) model.eval() A : List[Any] =model( __lowercase , attention_mask=__lowercase , start_positions=__lowercase , end_positions=__lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int ) -> str: A : Any =self.num_labels A : List[str] =DistilBertForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() A : Union[str, Any] =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : str ) -> int: A : Optional[Any] =self.num_labels A : Optional[int] =DistilBertForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() A : Any =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self : str , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int ) -> str: A : Dict =self.num_choices A : Dict =DistilBertForMultipleChoice(config=__lowercase ) model.to(__lowercase ) model.eval() A : Optional[int] =input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A : List[Any] =input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A : List[str] =model( __lowercase , attention_mask=__lowercase , labels=__lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Union[str, Any]: A : Any =self.prepare_config_and_inputs() (A) : int =config_and_inputs A : str ={"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' lowercase : Optional[Any] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) lowercase : Union[str, Any] = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) lowercase : Tuple = True lowercase : Tuple = True lowercase : Any = True lowercase : Optional[Any] = True def SCREAMING_SNAKE_CASE_ ( self : List[str] ) -> str: A : Dict =DistilBertModelTester(self ) A : Optional[int] =ConfigTester(self , config_class=__lowercase , dim=37 ) def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Optional[Any]: self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Any: A : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> Tuple: A : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : int ) -> Any: A : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : Any ) -> List[Any]: A : str =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> int: A : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Optional[int]: A : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*__lowercase ) @slow def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Dict: for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A : List[Any] =DistilBertModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @slow @require_torch_gpu def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> int: A : Dict =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return A : List[str] =True A : Dict =model_class(config=__lowercase ) A : List[Any] =self._prepare_for_class(__lowercase , __lowercase ) A : Dict =torch.jit.trace( __lowercase , (inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowercase , os.path.join(__lowercase , 'traced_model.pt' ) ) A : int =torch.jit.load(os.path.join(__lowercase , 'traced_model.pt' ) , map_location=__lowercase ) loaded(inputs_dict['input_ids'].to(__lowercase ) , inputs_dict['attention_mask'].to(__lowercase ) ) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Union[str, Any]: A : Tuple =DistilBertModel.from_pretrained('distilbert-base-uncased' ) A : Optional[int] =torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) A : List[str] =torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A : List[str] =model(__lowercase , attention_mask=__lowercase )[0] A : Dict =torch.Size((1, 11, 7_68) ) self.assertEqual(output.shape , __lowercase ) A : Dict =torch.tensor( [[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowercase , atol=1e-4 ) )
305
from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class a ( lowercase__ ): """simple docstring""" a : Optional[Any] = 'openai-gpt' a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __lowercase : Tuple=40478 , __lowercase : Tuple=512 , __lowercase : int=768 , __lowercase : Dict=12 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[Any]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Optional[Any]=0.1 , __lowercase : Tuple=0.1 , __lowercase : Dict=1e-5 , __lowercase : Any=0.02 , __lowercase : List[str]="cls_index" , __lowercase : str=True , __lowercase : Dict=None , __lowercase : str=True , __lowercase : List[str]=0.1 , **__lowercase : List[Any] , ) -> List[Any]: __UpperCAmelCase : Optional[int] = vocab_size __UpperCAmelCase : Optional[Any] = n_positions __UpperCAmelCase : Optional[int] = n_embd __UpperCAmelCase : str = n_layer __UpperCAmelCase : Any = n_head __UpperCAmelCase : Tuple = afn __UpperCAmelCase : Any = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : str = attn_pdrop __UpperCAmelCase : str = layer_norm_epsilon __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : Optional[int] = summary_type __UpperCAmelCase : Optional[Any] = summary_use_proj __UpperCAmelCase : List[Any] = summary_activation __UpperCAmelCase : Union[str, Any] = summary_first_dropout __UpperCAmelCase : Dict = summary_proj_to_labels super().__init__(**__lowercase )
63
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a : int = logging.get_logger(__name__) __a : Union[str, Any] = { "google/mobilenet_v2_1.4_224": "https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json", "google/mobilenet_v2_1.0_224": "https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json", "google/mobilenet_v2_0.75_160": "https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json", "google/mobilenet_v2_0.35_96": "https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json", # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class __lowercase ( lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mobilenet_v2' def __init__( self : Tuple , UpperCamelCase_ : List[Any]=3 , UpperCamelCase_ : List[str]=224 , UpperCamelCase_ : Tuple=1.0 , UpperCamelCase_ : Union[str, Any]=8 , UpperCamelCase_ : Tuple=8 , UpperCamelCase_ : List[Any]=6 , UpperCamelCase_ : Optional[int]=32 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Any=True , UpperCamelCase_ : List[Any]="relu6" , UpperCamelCase_ : Tuple=True , UpperCamelCase_ : str=0.8 , UpperCamelCase_ : List[str]=0.02 , UpperCamelCase_ : Union[str, Any]=0.001 , UpperCamelCase_ : Union[str, Any]=255 , **UpperCamelCase_ : str , ): """simple docstring""" super().__init__(**__lowercase ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) __A = num_channels __A = image_size __A = depth_multiplier __A = depth_divisible_by __A = min_depth __A = expand_ratio __A = output_stride __A = first_layer_is_expansion __A = finegrained_output __A = hidden_act __A = tf_padding __A = classifier_dropout_prob __A = initializer_range __A = layer_norm_eps __A = semantic_loss_ignore_index class __lowercase ( lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = version.parse("1.11" ) @property def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" return 1e-4
637
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = KandinskyVaaInpaintPipeline a : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] a : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] a : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] a : List[Any] = False @property def UpperCAmelCase ( self : int ) -> Dict: return 32 @property def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: return 32 @property def UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def UpperCAmelCase ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return 100 @property def UpperCAmelCase ( self : Dict ) -> Any: torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } __UpperCAmelCase : int = UNetaDConditionModel(**__lowercase ) return model @property def UpperCAmelCase ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self : Any ) -> List[Any]: __UpperCAmelCase : List[str] = self.dummy_unet __UpperCAmelCase : List[str] = self.dummy_movq __UpperCAmelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowercase , ) __UpperCAmelCase : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCAmelCase ( self : str , __lowercase : Tuple , __lowercase : List[str]=0 ) -> Optional[Any]: __UpperCAmelCase : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(__lowercase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask __UpperCAmelCase : Union[str, Any] = np.ones((64, 64) , dtype=np.floataa ) __UpperCAmelCase : List[str] = 0 if str(__lowercase ).startswith("""mps""" ): __UpperCAmelCase : List[str] = torch.manual_seed(__lowercase ) else: __UpperCAmelCase : Optional[int] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __UpperCAmelCase : Optional[Any] = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = """cpu""" __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**__lowercase ) __UpperCAmelCase : Tuple = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(__lowercase ) ) __UpperCAmelCase : Tuple = output.images __UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] __UpperCAmelCase : str = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __UpperCAmelCase : Optional[Any] = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def UpperCAmelCase ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) __UpperCAmelCase : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) __UpperCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa ) __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = """a hat""" __UpperCAmelCase : str = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) __UpperCAmelCase : Any = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) __UpperCAmelCase : int = pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() __UpperCAmelCase : Optional[int] = pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
63
0
"""simple docstring""" import functools def A_ ( __lowercase , __lowercase ): # Validation if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not all(isinstance(__lowerCamelCase , __lowerCamelCase ) for day in days ): raise ValueError('The parameter days should be a list of integers' ) if len(__lowerCamelCase ) != 3 or not all(isinstance(__lowerCamelCase , __lowerCamelCase ) for cost in costs ): raise ValueError('The parameter costs should be a list of three integers' ) if len(__lowerCamelCase ) == 0: return 0 if min(__lowerCamelCase ) <= 0: raise ValueError('All days elements should be greater than 0' ) if max(__lowerCamelCase ) >= 3_66: raise ValueError('All days elements should be less than 366' ) UpperCamelCase_ : List[str] =set(__lowerCamelCase ) @functools.cache def dynamic_programming(__lowercase ) -> int: if index > 3_65: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 30 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
357
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a : List[Any] = True except ImportError: a : str = False try: from torch.hub import _get_torch_home a : List[Any] = _get_torch_home() except ImportError: a : int = os.path.expanduser( os.getenv("TORCH_HOME", os.path.join(os.getenv("XDG_CACHE_HOME", "~/.cache"), "torch")) ) a : Optional[Any] = os.path.join(torch_cache_home, "transformers") a : Optional[Any] = "https://cdn.huggingface.co" a : List[str] = "https://s3.amazonaws.com/models.huggingface.co/bert" a : Any = "/".join(str(Path(__file__).resolve()).split("/")[:-1]) a : Optional[int] = os.path.join(PATH, "config.yaml") a : Dict = os.path.join(PATH, "attributes.txt") a : Tuple = os.path.join(PATH, "objects.txt") a : Dict = os.getenv("PYTORCH_PRETRAINED_BERT_CACHE", default_cache_path) a : Dict = os.getenv("PYTORCH_TRANSFORMERS_CACHE", PYTORCH_PRETRAINED_BERT_CACHE) a : Optional[int] = os.getenv("TRANSFORMERS_CACHE", PYTORCH_TRANSFORMERS_CACHE) a : Any = "pytorch_model.bin" a : int = "config.yaml" def lowerCamelCase__ ( __lowerCamelCase : str=OBJECTS , __lowerCamelCase : Union[str, Any]=ATTRIBUTES ): __UpperCAmelCase : Union[str, Any] = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(""",""" )[0].lower().strip() ) __UpperCAmelCase : Dict = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(""",""" )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : List[str] = OrderedDict() with open(__lowerCamelCase , """rb""" ) as f: __UpperCAmelCase : int = pkl.load(__lowerCamelCase )["""model"""] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : List[Any] = ckp.pop(__lowerCamelCase ) if isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : Union[str, Any] = torch.tensor(__lowerCamelCase ) else: assert isinstance(__lowerCamelCase , torch.tensor ), type(__lowerCamelCase ) __UpperCAmelCase : List[str] = v return r class a : """simple docstring""" a : Dict = {} def __init__( self : Dict , __lowercase : dict , __lowercase : str = "root" , __lowercase : Any=0 ) -> Dict: __UpperCAmelCase : List[str] = name __UpperCAmelCase : str = level __UpperCAmelCase : int = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : List[str] = copy.deepcopy(__lowercase ) __UpperCAmelCase : Dict = copy.deepcopy(__lowercase ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : Union[str, Any] = Config(__lowercase , name=__lowercase , level=level + 1 ) __UpperCAmelCase : Union[str, Any] = v setattr(self , __lowercase , __lowercase ) __UpperCAmelCase : Any = d def __repr__( self : Optional[Any] ) -> Optional[int]: return str(list((self._pointer.keys()) ) ) def __setattr__( self : List[str] , __lowercase : List[str] , __lowercase : Tuple ) -> int: __UpperCAmelCase : int = val __UpperCAmelCase : List[str] = val __UpperCAmelCase : Union[str, Any] = key.split(""".""" ) __UpperCAmelCase : List[Any] = len(__lowercase ) - 1 __UpperCAmelCase : List[Any] = self._pointer if len(__lowercase ) > 1: for i, l in enumerate(__lowercase ): if hasattr(self , __lowercase ) and isinstance(getattr(self , __lowercase ) , __lowercase ): setattr(getattr(self , __lowercase ) , """.""".join(levels[i:] ) , __lowercase ) if l == last_level: __UpperCAmelCase : Union[str, Any] = val else: __UpperCAmelCase : Union[str, Any] = pointer[l] def UpperCAmelCase ( self : Tuple ) -> Optional[int]: return self._pointer def UpperCAmelCase ( self : str , __lowercase : Optional[int] , __lowercase : Any ) -> Optional[int]: with open(f"""{file_name}""" , """w""" ) as stream: dump(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] ) -> Any: with open(f"""{file_name}""" , """w""" ) as stream: json.dump(__lowercase , __lowercase ) @staticmethod def UpperCAmelCase ( __lowercase : List[Any] ) -> Optional[Any]: with open(__lowercase ) as stream: __UpperCAmelCase : Any = load(__lowercase , Loader=__lowercase ) return data def __str__( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = """ """ if self._name != "root": __UpperCAmelCase : Optional[Any] = f"""{t * (self._level-1)}{self._name}:\n""" else: __UpperCAmelCase : List[Any] = """""" __UpperCAmelCase : Optional[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__lowercase , __lowercase ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__lowercase ).__name__})\n""" __UpperCAmelCase : int = level return r[:-1] @classmethod def UpperCAmelCase ( cls : List[str] , __lowercase : str , **__lowercase : Any ) -> Any: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) return cls(__lowercase ) @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : str , **__lowercase : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase : int = kwargs.pop("""cache_dir""" , __lowercase ) __UpperCAmelCase : int = kwargs.pop("""force_download""" , __lowercase ) __UpperCAmelCase : str = kwargs.pop("""resume_download""" , __lowercase ) __UpperCAmelCase : Dict = kwargs.pop("""proxies""" , __lowercase ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""local_files_only""" , __lowercase ) if os.path.isdir(__lowercase ): __UpperCAmelCase : List[Any] = os.path.join(__lowercase , __lowercase ) elif os.path.isfile(__lowercase ) or is_remote_url(__lowercase ): __UpperCAmelCase : Tuple = pretrained_model_name_or_path else: __UpperCAmelCase : Optional[int] = hf_bucket_url(__lowercase , filename=__lowercase , use_cdn=__lowercase ) try: # Load from URL or cache if already cached __UpperCAmelCase : Optional[int] = cached_path( __lowercase , cache_dir=__lowercase , force_download=__lowercase , proxies=__lowercase , resume_download=__lowercase , local_files_only=__lowercase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : Optional[int] = Config.load_yaml(__lowercase ) except EnvironmentError: __UpperCAmelCase : str = """Can't load config for""" raise EnvironmentError(__lowercase ) if resolved_config_file == config_file: print("""loading configuration file from path""" ) else: print("""loading configuration file cache""" ) return Config.load_yaml(__lowercase ), kwargs def lowerCamelCase__ ( __lowerCamelCase : Dict ): __UpperCAmelCase : Optional[int] = torch.load("""dump.pt""" , map_location=in_tensor.device ) __UpperCAmelCase : Tuple = in_tensor.numpy() __UpperCAmelCase : Optional[int] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception("""tensors are all good""" ) # Hugging face functions below def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Tuple = urlparse(__lowerCamelCase ) return parsed.scheme in ("http", "https") def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : int=True ): __UpperCAmelCase : int = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : Optional[int] = """/""" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=None , ): __UpperCAmelCase : Optional[int] = """python/{}""".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + "; ".join("""{}/{}""".format(__lowerCamelCase , __lowerCamelCase ) for k, v in user_agent.items() ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + user_agent __UpperCAmelCase : List[str] = {"""user-agent""": ua} if resume_size > 0: __UpperCAmelCase : Union[str, Any] = """bytes=%d-""" % (resume_size,) __UpperCAmelCase : Union[str, Any] = requests.get(__lowerCamelCase , stream=__lowerCamelCase , proxies=__lowerCamelCase , headers=__lowerCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : List[str] = response.headers.get("""Content-Length""" ) __UpperCAmelCase : str = resume_size + int(__lowerCamelCase ) if content_length is not None else None __UpperCAmelCase : List[Any] = tqdm( unit="""B""" , unit_scale=__lowerCamelCase , total=__lowerCamelCase , initial=__lowerCamelCase , desc="""Downloading""" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(__lowerCamelCase ) ) temp_file.write(__lowerCamelCase ) progress.close() def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=10 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=None , __lowerCamelCase : List[str]=False , ): if cache_dir is None: __UpperCAmelCase : Optional[Any] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[str] = str(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) __UpperCAmelCase : List[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(__lowerCamelCase , allow_redirects=__lowerCamelCase , proxies=__lowerCamelCase , timeout=__lowerCamelCase ) if response.status_code == 200: __UpperCAmelCase : Dict = response.headers.get("""ETag""" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : List[str] = url_to_filename(__lowerCamelCase , __lowerCamelCase ) # get cache path to put the file __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , __lowerCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(__lowerCamelCase ): return cache_path else: __UpperCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(__lowerCamelCase ) , filename + """.*""" ) if not file.endswith(""".json""" ) and not file.endswith(""".lock""" ) ] if len(__lowerCamelCase ) > 0: return os.path.join(__lowerCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( """Cannot find the requested files in the cached path and outgoing traffic has been""" """ disabled. To enable model look-ups and downloads online, set 'local_files_only'""" """ to False.""" ) return None # From now on, etag is not None. if os.path.exists(__lowerCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : str = cache_path + """.lock""" with FileLock(__lowerCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(__lowerCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : int = cache_path + """.incomplete""" @contextmanager def _resumable_file_manager(): with open(__lowerCamelCase , """a+b""" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(__lowerCamelCase ): __UpperCAmelCase : List[Any] = os.stat(__lowerCamelCase ).st_size else: __UpperCAmelCase : List[Any] = 0 else: __UpperCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=__lowerCamelCase , delete=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( """%s not found in cache or force_download set to True, downloading to %s""" , __lowerCamelCase , temp_file.name , ) http_get( __lowerCamelCase , __lowerCamelCase , proxies=__lowerCamelCase , resume_size=__lowerCamelCase , user_agent=__lowerCamelCase , ) os.replace(temp_file.name , __lowerCamelCase ) __UpperCAmelCase : Any = {"""url""": url, """etag""": etag} __UpperCAmelCase : Union[str, Any] = cache_path + """.json""" with open(__lowerCamelCase , """w""" ) as meta_file: json.dump(__lowerCamelCase , __lowerCamelCase ) return cache_path def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any]=None ): __UpperCAmelCase : Tuple = url.encode("""utf-8""" ) __UpperCAmelCase : Optional[Any] = shaaaa(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = url_hash.hexdigest() if etag: __UpperCAmelCase : int = etag.encode("""utf-8""" ) __UpperCAmelCase : List[str] = shaaaa(__lowerCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(""".h5""" ): filename += ".h5" return filename def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : int=None , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=False , ): if cache_dir is None: __UpperCAmelCase : List[str] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Any = str(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) if is_remote_url(__lowerCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : Tuple = get_from_cache( __lowerCamelCase , cache_dir=__lowerCamelCase , force_download=__lowerCamelCase , proxies=__lowerCamelCase , resume_download=__lowerCamelCase , user_agent=__lowerCamelCase , local_files_only=__lowerCamelCase , ) elif os.path.exists(__lowerCamelCase ): # File, and it exists. __UpperCAmelCase : Tuple = url_or_filename elif urlparse(__lowerCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("""file {} not found""".format(__lowerCamelCase ) ) else: # Something unknown raise ValueError("""unable to parse {} as a URL or as a local path""".format(__lowerCamelCase ) ) if extract_compressed_file: if not is_zipfile(__lowerCamelCase ) and not tarfile.is_tarfile(__lowerCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : int = os.path.split(__lowerCamelCase ) __UpperCAmelCase : Any = output_file.replace(""".""" , """-""" ) + """-extracted""" __UpperCAmelCase : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isdir(__lowerCamelCase ) and os.listdir(__lowerCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : str = output_path + """.lock""" with FileLock(__lowerCamelCase ): shutil.rmtree(__lowerCamelCase , ignore_errors=__lowerCamelCase ) os.makedirs(__lowerCamelCase ) if is_zipfile(__lowerCamelCase ): with ZipFile(__lowerCamelCase , """r""" ) as zip_file: zip_file.extractall(__lowerCamelCase ) zip_file.close() elif tarfile.is_tarfile(__lowerCamelCase ): __UpperCAmelCase : Any = tarfile.open(__lowerCamelCase ) tar_file.extractall(__lowerCamelCase ) tar_file.close() else: raise EnvironmentError("""Archive format of {} could not be identified""".format(__lowerCamelCase ) ) return output_path_extracted return output_path def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : int="," ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): with open(__lowerCamelCase ) as f: __UpperCAmelCase : List[Any] = eval(f.read() ) else: __UpperCAmelCase : List[str] = requests.get(__lowerCamelCase ) try: __UpperCAmelCase : int = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : str = eval(__lowerCamelCase ) except Exception: __UpperCAmelCase : List[Any] = data.split("""\n""" ) req.close() return data def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : Optional[int] = requests.get(__lowerCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCamelCase__ ( __lowerCamelCase : str ): __UpperCAmelCase : int = url.split("""/""" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(__lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as stream: __UpperCAmelCase : List[str] = pkl.load(__lowerCamelCase ) __UpperCAmelCase : Dict = weights.pop("""model""" ) __UpperCAmelCase : Union[str, Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(__lowerCamelCase ) if "running_var" in k: __UpperCAmelCase : Optional[int] = torch.tensor([0] ) __UpperCAmelCase : Tuple = k.replace("""running_var""" , """num_batches_tracked""" ) __UpperCAmelCase : Any = zero return new def lowerCamelCase__ ( ): print(f"""{os.path.abspath(os.path.join(__lowerCamelCase , os.pardir ) )}/demo.ipynb""" ) def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]="RGB" ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): __UpperCAmelCase : List[str] = cva.imread(__lowerCamelCase ) else: __UpperCAmelCase : int = get_image_from_url(__lowerCamelCase ) assert img is not None, f"""could not connect to: {im}""" __UpperCAmelCase : Any = cva.cvtColor(__lowerCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Optional[int] = img[:, :, ::-1] return img def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int=1 ): return (images[i : i + batch] for i in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ))
63
0
import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__(lowercase__, unittest.TestCase ): """simple docstring""" _A : List[Any] = AudioLDMPipeline _A : Optional[Any] = TEXT_TO_AUDIO_PARAMS _A : Dict = TEXT_TO_AUDIO_BATCH_PARAMS _A : Optional[int] = frozenset( [ '''num_inference_steps''', '''num_waveforms_per_prompt''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def UpperCamelCase__ ( self ) -> List[str]: torch.manual_seed(0 ) a_ : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=(32, 64) , class_embed_type="""simple_projection""" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=__lowercase , ) a_ : Optional[int] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0 ) a_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) a_ : Optional[Any] = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , projection_dim=32 , ) a_ : Optional[int] = ClapTextModelWithProjection(__lowercase ) a_ : str = RobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-roberta""" , model_max_length=77 ) a_ : Dict = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=16_000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=__lowercase , ) a_ : int = SpeechTaHifiGan(__lowercase ) a_ : Tuple = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """vocoder""": vocoder, } return components def UpperCamelCase__ ( self , _lowercase , _lowercase=0 ) -> List[str]: if str(__lowercase ).startswith("""mps""" ): a_ : Dict = torch.manual_seed(__lowercase ) else: a_ : Tuple = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) a_ : Tuple = { """prompt""": """A hammer hitting a wooden surface""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, } return inputs def UpperCamelCase__ ( self ) -> Union[str, Any]: a_ : Tuple = """cpu""" # ensure determinism for the device-dependent torch.Generator a_ : Dict = self.get_dummy_components() a_ : List[Any] = AudioLDMPipeline(**__lowercase ) a_ : Tuple = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : List[Any] = self.get_dummy_inputs(__lowercase ) a_ : Union[str, Any] = audioldm_pipe(**__lowercase ) a_ : Union[str, Any] = output.audios[0] assert audio.ndim == 1 assert len(__lowercase ) == 256 a_ : str = audio[:10] a_ : List[Any] = np.array( [-0.0_0_5_0, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_3, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_3] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ) -> Optional[int]: a_ : List[str] = self.get_dummy_components() a_ : Any = AudioLDMPipeline(**__lowercase ) a_ : Tuple = audioldm_pipe.to(__lowercase ) a_ : str = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : Tuple = self.get_dummy_inputs(__lowercase ) a_ : Dict = 3 * [inputs["""prompt"""]] # forward a_ : Union[str, Any] = audioldm_pipe(**__lowercase ) a_ : int = output.audios[0] a_ : List[str] = self.get_dummy_inputs(__lowercase ) a_ : Any = 3 * [inputs.pop("""prompt""" )] a_ : Tuple = audioldm_pipe.tokenizer( __lowercase , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=__lowercase , return_tensors="""pt""" , ) a_ : Optional[Any] = text_inputs["""input_ids"""].to(__lowercase ) a_ : int = audioldm_pipe.text_encoder( __lowercase , ) a_ : Dict = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state a_ : Tuple = F.normalize(__lowercase , dim=-1 ) a_ : Tuple = prompt_embeds # forward a_ : Dict = audioldm_pipe(**__lowercase ) a_ : str = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def UpperCamelCase__ ( self ) -> str: a_ : Tuple = self.get_dummy_components() a_ : Any = AudioLDMPipeline(**__lowercase ) a_ : Dict = audioldm_pipe.to(__lowercase ) a_ : int = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : Union[str, Any] = self.get_dummy_inputs(__lowercase ) a_ : Optional[Any] = 3 * ["""this is a negative prompt"""] a_ : Optional[Any] = negative_prompt a_ : Tuple = 3 * [inputs["""prompt"""]] # forward a_ : int = audioldm_pipe(**__lowercase ) a_ : Any = output.audios[0] a_ : List[Any] = self.get_dummy_inputs(__lowercase ) a_ : Tuple = 3 * [inputs.pop("""prompt""" )] a_ : List[Any] = [] for p in [prompt, negative_prompt]: a_ : List[str] = audioldm_pipe.tokenizer( __lowercase , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=__lowercase , return_tensors="""pt""" , ) a_ : Union[str, Any] = text_inputs["""input_ids"""].to(__lowercase ) a_ : Optional[Any] = audioldm_pipe.text_encoder( __lowercase , ) a_ : Tuple = text_embeds.text_embeds # additional L_2 normalization over each hidden-state a_ : Any = F.normalize(__lowercase , dim=-1 ) embeds.append(__lowercase ) a_ : Optional[int] = embeds # forward a_ : str = audioldm_pipe(**__lowercase ) a_ : str = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def UpperCamelCase__ ( self ) -> Tuple: a_ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator a_ : List[Any] = self.get_dummy_components() a_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=__lowercase ) a_ : Tuple = AudioLDMPipeline(**__lowercase ) a_ : str = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : Union[str, Any] = self.get_dummy_inputs(__lowercase ) a_ : Optional[Any] = """egg cracking""" a_ : Optional[Any] = audioldm_pipe(**__lowercase , negative_prompt=__lowercase ) a_ : Tuple = output.audios[0] assert audio.ndim == 1 assert len(__lowercase ) == 256 a_ : Union[str, Any] = audio[:10] a_ : int = np.array( [-0.0_0_5_1, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_4, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_2] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ) -> Any: a_ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator a_ : List[Any] = self.get_dummy_components() a_ : str = PNDMScheduler(skip_prk_steps=__lowercase ) a_ : Tuple = AudioLDMPipeline(**__lowercase ) a_ : Tuple = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : str = """A hammer hitting a wooden surface""" # test num_waveforms_per_prompt=1 (default) a_ : Union[str, Any] = audioldm_pipe(__lowercase , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts a_ : Optional[Any] = 2 a_ : int = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt a_ : int = 2 a_ : str = audioldm_pipe(__lowercase , num_inference_steps=2 , num_waveforms_per_prompt=__lowercase ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts a_ : Any = 2 a_ : Tuple = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=__lowercase ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def UpperCamelCase__ ( self ) -> str: a_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator a_ : Tuple = self.get_dummy_components() a_ : int = AudioLDMPipeline(**__lowercase ) a_ : Dict = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : List[str] = audioldm_pipe.vocoder.config.sampling_rate a_ : Union[str, Any] = self.get_dummy_inputs(__lowercase ) a_ : Optional[Any] = audioldm_pipe(audio_length_in_s=0.0_1_6 , **__lowercase ) a_ : Tuple = output.audios[0] assert audio.ndim == 1 assert len(__lowercase ) / vocoder_sampling_rate == 0.0_1_6 a_ : Optional[Any] = audioldm_pipe(audio_length_in_s=0.0_3_2 , **__lowercase ) a_ : Dict = output.audios[0] assert audio.ndim == 1 assert len(__lowercase ) / vocoder_sampling_rate == 0.0_3_2 def UpperCamelCase__ ( self ) -> List[Any]: a_ : List[Any] = self.get_dummy_components() a_ : Any = AudioLDMPipeline(**__lowercase ) a_ : Dict = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : List[str] = ["""hey"""] a_ : Dict = audioldm_pipe(__lowercase , num_inference_steps=1 ) a_ : Tuple = output.audios.shape assert audio_shape == (1, 256) a_ : Optional[Any] = audioldm_pipe.vocoder.config config.model_in_dim *= 2 a_ : List[Any] = SpeechTaHifiGan(__lowercase ).to(__lowercase ) a_ : Dict = audioldm_pipe(__lowercase , num_inference_steps=1 ) a_ : int = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def UpperCamelCase__ ( self ) -> Optional[int]: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase ) def UpperCamelCase__ ( self ) -> Any: self._test_inference_batch_single_identical(test_mean_pixel_difference=__lowercase ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase__ ( self ) -> Optional[Any]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase ) @slow class A__(unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , _lowercase , _lowercase="cpu" , _lowercase=torch.floataa , _lowercase=0 ) -> Dict: a_ : int = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) a_ : Dict = np.random.RandomState(__lowercase ).standard_normal((1, 8, 128, 16) ) a_ : Optional[Any] = torch.from_numpy(__lowercase ).to(device=__lowercase , dtype=__lowercase ) a_ : int = { """prompt""": """A hammer hitting a wooden surface""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 2.5, } return inputs def UpperCamelCase__ ( self ) -> List[str]: a_ : Any = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) a_ : Union[str, Any] = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : Tuple = self.get_inputs(__lowercase ) a_ : str = 25 a_ : Optional[int] = audioldm_pipe(**__lowercase ).audios[0] assert audio.ndim == 1 assert len(__lowercase ) == 81_920 a_ : Dict = audio[77_230:77_240] a_ : Optional[Any] = np.array( [-0.4_8_8_4, -0.4_6_0_7, 0.0_0_2_3, 0.5_0_0_7, 0.5_8_9_6, 0.5_1_5_1, 0.3_8_1_3, -0.0_2_0_8, -0.3_6_8_7, -0.4_3_1_5] ) a_ : Optional[Any] = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def UpperCamelCase__ ( self ) -> Tuple: a_ : Optional[Any] = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) a_ : Any = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) a_ : int = audioldm_pipe.to(__lowercase ) audioldm_pipe.set_progress_bar_config(disable=__lowercase ) a_ : List[Any] = self.get_inputs(__lowercase ) a_ : Optional[int] = audioldm_pipe(**__lowercase ).audios[0] assert audio.ndim == 1 assert len(__lowercase ) == 81_920 a_ : int = audio[27_780:27_790] a_ : Optional[Any] = np.array([-0.2_1_3_1, -0.0_8_7_3, -0.0_1_2_4, -0.0_1_8_9, 0.0_5_6_9, 0.1_3_7_3, 0.1_8_8_3, 0.2_8_8_6, 0.3_2_9_7, 0.2_2_1_2] ) a_ : Dict = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
540
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Any=13 , __lowercase : Optional[int]=7 , __lowercase : str=True , __lowercase : Optional[Any]=True , __lowercase : int=True , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : int=5 , __lowercase : Tuple=4 , __lowercase : str=37 , __lowercase : Optional[int]="gelu" , __lowercase : Tuple=0.1 , __lowercase : str=0.1 , __lowercase : Dict=512 , __lowercase : List[Any]=16 , __lowercase : Dict=2 , __lowercase : Union[str, Any]=0.02 , __lowercase : Dict=4 , ) -> int: __UpperCAmelCase : Dict = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : str = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Dict = use_token_type_ids __UpperCAmelCase : Optional[int] = use_labels __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : str = type_sequence_label_size __UpperCAmelCase : List[str] = initializer_range __UpperCAmelCase : Tuple = num_choices def UpperCAmelCase ( self : Dict ) -> Tuple: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : List[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase ( self : Tuple ) -> List[Any]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase ( self : Any ) -> List[str]: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[int] = config_and_inputs __UpperCAmelCase : int = True __UpperCAmelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = True a : List[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : List[str] = FlaxRobertaModelTester(self ) @slow def UpperCAmelCase ( self : str ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase : Union[str, Any] = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowercase ) __UpperCAmelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowercase )
63
0
'''simple docstring''' import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging __magic_name__ : List[str] =logging.get_logger(__name__) class UpperCamelCase_ ( lowercase__ ): """simple docstring""" UpperCAmelCase__ : Tuple = CLIPConfig UpperCAmelCase__ : Optional[Any] = ['CLIPEncoderLayer'] def __init__( self : List[str] , _lowerCamelCase : CLIPConfig ) -> List[Any]: super().__init__(__lowercase ) __magic_name__ = CLIPVisionModelWithProjection(config.vision_config ) __magic_name__ = nn.Linear(config.vision_config.projection_dim , 1 ) __magic_name__ = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def __A ( self : Any , _lowerCamelCase : List[str] , _lowerCamelCase : Dict , _lowerCamelCase : Any=0.5 , _lowerCamelCase : Dict=0.5 ) -> List[Any]: __magic_name__ = self.vision_model(__lowercase )[0] __magic_name__ = self.p_head(__lowercase ) __magic_name__ = nsfw_detected.flatten() __magic_name__ = nsfw_detected > p_threshold __magic_name__ = nsfw_detected.tolist() if any(__lowercase ): logger.warning( "Potential NSFW content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, nsfw_detected_ in enumerate(__lowercase ): if nsfw_detected_: __magic_name__ = np.zeros(images[idx].shape ) __magic_name__ = self.w_head(__lowercase ) __magic_name__ = watermark_detected.flatten() __magic_name__ = watermark_detected > w_threshold __magic_name__ = watermark_detected.tolist() if any(__lowercase ): logger.warning( "Potential watermarked content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, watermark_detected_ in enumerate(__lowercase ): if watermark_detected_: __magic_name__ = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
664
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging a : Optional[int] = logging.get_logger(__name__) class a ( lowercase__ ): """simple docstring""" a : Tuple = 'linear' a : int = 'cosine' a : Optional[Any] = 'cosine_with_restarts' a : Dict = 'polynomial' a : Tuple = 'constant' a : Dict = 'constant_with_warmup' a : Any = 'piecewise_constant' def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int = -1 ): return LambdaLR(__lowerCamelCase , lambda __lowerCamelCase : 1 , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : str , __lowerCamelCase : int = -1 ): __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Tuple = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __UpperCAmelCase , __UpperCAmelCase : List[str] = rule_str.split(""":""" ) __UpperCAmelCase : Any = int(__lowerCamelCase ) __UpperCAmelCase : List[str] = float(__lowerCamelCase ) __UpperCAmelCase : int = value __UpperCAmelCase : Any = float(rule_list[-1] ) def create_rules_function(__lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): def rule_func(__lowerCamelCase : int ) -> float: __UpperCAmelCase : Tuple = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=-1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float = 0.5 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Dict ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Tuple = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int = 1 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Union[str, Any] ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Union[str, Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=1E-7 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : int=-1 ): __UpperCAmelCase : Tuple = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(f"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __UpperCAmelCase : Optional[Any] = lr_init - lr_end __UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps __UpperCAmelCase : int = 1 - (current_step - num_warmup_steps) / decay_steps __UpperCAmelCase : Optional[int] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a : int = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCamelCase__ ( __lowerCamelCase : Union[str, SchedulerType] , __lowerCamelCase : Optimizer , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 1 , __lowerCamelCase : float = 1.0 , __lowerCamelCase : int = -1 , ): __UpperCAmelCase : Union[str, Any] = SchedulerType(__lowerCamelCase ) __UpperCAmelCase : int = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
63
0
_UpperCamelCase = [ "Audio", "Array2D", "Array3D", "Array4D", "Array5D", "ClassLabel", "Features", "Sequence", "Value", "Image", "Translation", "TranslationVariableLanguages", ] from .audio import Audio from .features import ArrayaD, ArrayaD, ArrayaD, ArrayaD, ClassLabel, Features, Sequence, Value from .image import Image from .translation import Translation, TranslationVariableLanguages
146
from math import pi, sqrt def lowerCamelCase__ ( __lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCamelCase__ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() a : Optional[int] = 1.0 while num: a : List[str] = float(input("Gamma of: ")) print(f"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
63
0
'''simple docstring''' from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Tuple ,lowercase__ : Any ,lowercase__ : Union[str, Any]=3 ,lowercase__ : str=3_2 ,lowercase__ : List[str]=3 ,lowercase__ : Union[str, Any]=1_0 ,lowercase__ : int=[1_0, 2_0, 3_0, 4_0] ,lowercase__ : Any=[1, 1, 2, 1] ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=True ,lowercase__ : str="relu" ,lowercase__ : int=3 ,lowercase__ : Any=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = embeddings_size __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = hidden_act __lowercase = num_labels __lowercase = scope __lowercase = len(__lowercase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return ResNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ): __lowercase = TFResNetModel(config=__lowercase ) __lowercase = model(__lowercase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) ,) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : Any ): __lowercase = self.num_labels __lowercase = TFResNetForImageClassification(__lowercase ) __lowercase = model(__lowercase ,labels=__lowercase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.prepare_config_and_inputs() __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class lowercase_ (lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE : Union[str, Any] = ( {'feature-extraction': TFResNetModel, 'image-classification': TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = TFResNetModelTester(self ) __lowercase = ConfigTester(self ,config_class=__lowercase ,has_text_modality=__lowercase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : List[Any] ): return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(__lowercase ) __lowercase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def SCREAMING_SNAKE_CASE ( self : Any ): def check_hidden_states_output(lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Tuple ): __lowercase = model_class(__lowercase ) __lowercase = model(**self._prepare_for_class(__lowercase ,__lowercase ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(__lowercase ) ,expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 4, self.model_tester.image_size // 4] ,) __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: __lowercase = layer_type __lowercase = True check_hidden_states_output(__lowercase ,__lowercase ,__lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(__lowercase ,__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) @slow def SCREAMING_SNAKE_CASE ( self : List[Any] ): for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = TFResNetModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Any ): return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=__lowercase ,return_tensors='''tf''' ) # forward pass __lowercase = model(**__lowercase ) # verify the logits __lowercase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,__lowercase ) __lowercase = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() ,__lowercase ,atol=1e-4 ) )
41
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class a : """simple docstring""" a : int a : Node | None = None a : Node | None = None def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = Node(1 ) __UpperCAmelCase : int = Node(2 ) __UpperCAmelCase : Optional[Any] = Node(3 ) __UpperCAmelCase : Dict = Node(4 ) __UpperCAmelCase : Tuple = Node(5 ) return tree def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCamelCase__ ( __lowerCamelCase : Node | None ): __UpperCAmelCase : list[Any] = [] if root is None: return output __UpperCAmelCase : Tuple = deque([root] ) while process_queue: __UpperCAmelCase : Optional[Any] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None ): if root is None: return [] __UpperCAmelCase : list[Sequence[Node | None]] = [] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : int = height(__lowerCamelCase ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : Optional[int] = 0 return output def lowerCamelCase__ ( ): # Main function for testing. __UpperCAmelCase : List[Any] = make_tree() print(f"""In-order Traversal: {inorder(__lowerCamelCase )}""" ) print(f"""Pre-order Traversal: {preorder(__lowerCamelCase )}""" ) print(f"""Post-order Traversal: {postorder(__lowerCamelCase )}""" , """\n""" ) print(f"""Height of Tree: {height(__lowerCamelCase )}""" , """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(__lowerCamelCase ) , """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1 , height(__lowerCamelCase ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(__lowerCamelCase , level=__lowerCamelCase ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
283
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[int] = GPTSanJapaneseTokenizer a : Optional[Any] = False a : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off __UpperCAmelCase : Tuple = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __UpperCAmelCase : Dict = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __UpperCAmelCase : Dict = {"""unk_token""": """<unk>"""} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(__lowercase ) ) def UpperCAmelCase ( self : Tuple , **__lowercase : int ) -> Any: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def UpperCAmelCase ( self : str , __lowercase : Union[str, Any] ) -> Any: __UpperCAmelCase : Any = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __UpperCAmelCase : int = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase ( self : List[Any] , __lowercase : Optional[int] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase : int = self.get_input_output_texts(__lowercase ) __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : Dict = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) return text, ids def UpperCAmelCase ( self : int ) -> Optional[Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Dict ) -> Tuple: pass # TODO add if relevant def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : List[str] = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。 こんばんは、㔺界。""" __UpperCAmelCase : Dict = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids without special tokens __UpperCAmelCase : List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids with special tokens __UpperCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __UpperCAmelCase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : int = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : Tuple = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __UpperCAmelCase : int = """こんにちは、、、、世界。こんばんは、、、、世界。""" __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase ) __UpperCAmelCase : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : int ) -> Optional[int]: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : List[Any] = """こんにちは、世界。""" __UpperCAmelCase : Optional[int] = """こんばんは、㔺界。😀""" __UpperCAmelCase : List[Any] = """こんにちは、世界。こんばんは、世界。😀""" __UpperCAmelCase : List[str] = tokenizer.encode(prefix_text + input_text ) __UpperCAmelCase : List[Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __UpperCAmelCase : Any = tokenizer.encode(__lowercase , prefix_text=__lowercase ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowercase ) __UpperCAmelCase : Any = tokenizer.decode(__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。""" __UpperCAmelCase : List[Any] = """こんばんは、㔺界。😀""" __UpperCAmelCase : Union[str, Any] = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : int = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : List[Any] = [1] + [0] * (len_prefix + len_text + 1) __UpperCAmelCase : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] __UpperCAmelCase : List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCAmelCase : Union[str, Any] = tokenizer(prefix_text + input_text ).token_type_ids __UpperCAmelCase : Optional[Any] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __UpperCAmelCase : Tuple = tokenizer(__lowercase , prefix_text=__lowercase ).token_type_ids self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : List[str] ) -> int: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""あンいワ""" ) __UpperCAmelCase : Tuple = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertNotEqual(__lowercase , __lowercase ) self.assertNotEqual(__lowercase , __lowercase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self : List[Any] ) -> List[str]: __UpperCAmelCase : Any = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : List[Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __UpperCAmelCase : int = tokenizer(__lowercase , padding=__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.batch_encode_plus(__lowercase , padding=__lowercase ) # fmt: off __UpperCAmelCase : Optional[int] = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCAmelCase : Tuple = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCAmelCase : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __lowercase ) self.assertListEqual(x_token.token_type_ids , __lowercase ) self.assertListEqual(x_token.attention_mask , __lowercase ) self.assertListEqual(x_token_a.input_ids , __lowercase ) self.assertListEqual(x_token_a.token_type_ids , __lowercase ) self.assertListEqual(x_token_a.attention_mask , __lowercase ) def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase ( self : Any ) -> int: # tokenizer has no padding token pass
63
0
def __UpperCamelCase ( lowercase__ : int , lowercase__ : int , lowercase__ : int ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : int = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def __UpperCamelCase ( ) -> str: '''simple docstring''' print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
600
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a : Dict = logging.get_logger(__name__) @dataclass class a ( lowercase__ ): """simple docstring""" a : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : List[Any] , **__lowercase : Dict ) -> Tuple: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase : List[Any] = deprecated_arg[3:] setattr(self , __lowercase , not kwargs.pop(__lowercase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __UpperCAmelCase : str = kwargs.pop("""torchscript""" , self.torchscript ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) __UpperCAmelCase : Optional[Any] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**__lowercase ) a : bool = field(default=lowercase__ , metadata={'help': 'Trace the models using torchscript'} ) a : bool = field(default=lowercase__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) a : str = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def UpperCAmelCase ( self : Any ) -> Tuple["torch.device", int]: requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: __UpperCAmelCase : str = torch.device("""cpu""" ) __UpperCAmelCase : int = 0 elif is_torch_tpu_available(): __UpperCAmelCase : Tuple = xm.xla_device() __UpperCAmelCase : int = 0 else: __UpperCAmelCase : Dict = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) __UpperCAmelCase : Optional[int] = torch.cuda.device_count() return device, n_gpu @property def UpperCAmelCase ( self : Optional[Any] ) -> str: return is_torch_tpu_available() and self.tpu @property def UpperCAmelCase ( self : List[str] ) -> int: requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCAmelCase ( self : int ) -> "torch.device": requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def UpperCAmelCase ( self : int ) -> List[Any]: requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return self.n_gpu > 0
63
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { "openai/imagegpt-small": "", "openai/imagegpt-medium": "", "openai/imagegpt-large": "", } class UpperCamelCase_ ( lowercase__ ): lowerCamelCase_ = 'imagegpt' lowerCamelCase_ = ['past_key_values'] lowerCamelCase_ = { 'hidden_size': 'n_embd', 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self :Union[str, Any] , __A :Dict=512 + 1 , __A :int=32 * 32 , __A :List[str]=512 , __A :Dict=24 , __A :Optional[int]=8 , __A :Tuple=None , __A :Tuple="quick_gelu" , __A :int=0.1 , __A :Union[str, Any]=0.1 , __A :Optional[int]=0.1 , __A :List[Any]=1E-5 , __A :Optional[Any]=0.0_2 , __A :Union[str, Any]=True , __A :Optional[int]=True , __A :Optional[int]=False , __A :Union[str, Any]=False , __A :Any=False , **__A :Optional[Any] , ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = n_positions SCREAMING_SNAKE_CASE__ = n_embd SCREAMING_SNAKE_CASE__ = n_layer SCREAMING_SNAKE_CASE__ = n_head SCREAMING_SNAKE_CASE__ = n_inner SCREAMING_SNAKE_CASE__ = activation_function SCREAMING_SNAKE_CASE__ = resid_pdrop SCREAMING_SNAKE_CASE__ = embd_pdrop SCREAMING_SNAKE_CASE__ = attn_pdrop SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = scale_attn_weights SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = scale_attn_by_inverse_layer_idx SCREAMING_SNAKE_CASE__ = reorder_and_upcast_attn SCREAMING_SNAKE_CASE__ = tie_word_embeddings super().__init__(tie_word_embeddings=__lowercase , **__lowercase ) class UpperCamelCase_ ( lowercase__ ): @property def _snake_case ( self :Tuple ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ] ) def _snake_case ( self :Union[str, Any] , __A :"FeatureExtractionMixin" , __A :int = 1 , __A :int = -1 , __A :bool = False , __A :Optional["TensorType"] = None , __A :int = 3 , __A :int = 32 , __A :int = 32 , ) -> Mapping[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self._generate_dummy_images(__lowercase , __lowercase , __lowercase , __lowercase ) SCREAMING_SNAKE_CASE__ = dict(preprocessor(images=__lowercase , return_tensors=__lowercase ) ) return inputs
6
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __UpperCAmelCase : str = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Any ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __UpperCAmelCase : Any = features.copy() if features else default_expected_features __UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=__lowerCamelCase , cache_dir=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: __UpperCAmelCase : Dict = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): __UpperCAmelCase : Optional[int] = tmp_path / """cache""" __UpperCAmelCase : str = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() __UpperCAmelCase : Optional[int] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Dict = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): __UpperCAmelCase : int = tmp_path / """cache""" __UpperCAmelCase : int = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Any = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Optional[int] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() with pytest.raises(__lowerCamelCase ): SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
63
0
from __future__ import annotations from typing import TypedDict class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' A : str A : int def lowerCAmelCase__ ( _a : str ): if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(__lowerCamelCase ) )] def lowerCAmelCase__ ( _a : str ): if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) snake_case_ : List[Any] = all_rotations(__lowerCamelCase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation snake_case_ : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__lowerCamelCase ), } return response def lowerCAmelCase__ ( _a : str , _a : int ): if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: snake_case_ : Tuple = int(__lowerCamelCase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(__lowerCamelCase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) snake_case_ : Tuple = [""""""] * len(__lowerCamelCase ) for _ in range(len(__lowerCamelCase ) ): for i in range(len(__lowerCamelCase ) ): snake_case_ : Optional[Any] = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": lowercase : int = "Provide a string that I will generate its BWT transform: " lowercase : str = input(entry_msg).strip() lowercase : List[Any] = bwt_transform(s) print( F"""Burrows Wheeler transform for string '{s}' results """ F"""in '{result["bwt_string"]}'""" ) lowercase : Union[str, Any] = reverse_bwt(result['''bwt_string'''], result['''idx_original_string''']) print( F"""Reversing Burrows Wheeler transform for entry '{result["bwt_string"]}' """ F"""we get original string '{original_string}'""" )
568
from __future__ import annotations a : Optional[Any] = [True] * 1_000_001 a : Union[str, Any] = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): a : Optional[Any] = False i += 1 def lowerCamelCase__ ( __lowerCamelCase : int ): return seive[n] def lowerCamelCase__ ( __lowerCamelCase : int ): return any(digit in """02468""" for digit in str(__lowerCamelCase ) ) def lowerCamelCase__ ( __lowerCamelCase : int = 1000000 ): __UpperCAmelCase : Optional[Any] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__lowerCamelCase ) and not contains_an_even_digit(__lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) __UpperCAmelCase : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__lowerCamelCase ) )] if all(is_prime(__lowerCamelCase ) for i in list_nums ): result.append(__lowerCamelCase ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"""{len(find_circular_primes()) = }""")
63
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu _lowercase : Any =get_tests_dir() + "/test_data/fsmt/fsmt_val_data.json" with io.open(filename, '''r''', encoding='''utf-8''') as f: _lowercase : Union[str, Any] =json.load(f) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE_ ( self : int , SCREAMING_SNAKE_CASE__ : List[Any] ) -> Tuple: return FSMTTokenizer.from_pretrained(__lowercase ) def SCREAMING_SNAKE_CASE_ ( self : Any , SCREAMING_SNAKE_CASE__ : Any ) -> Any: A : Any =FSMTForConditionalGeneration.from_pretrained(__lowercase ).to(__lowercase ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['en-ru', 2_6.0], ['ru-en', 2_2.0], ['en-de', 2_2.0], ['de-en', 2_9.0], ] ) @slow def SCREAMING_SNAKE_CASE_ ( self : List[str] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[str]: # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality A : Dict =f'facebook/wmt19-{pair}' A : Union[str, Any] =self.get_tokenizer(__lowercase ) A : Any =self.get_model(__lowercase ) A : Optional[Any] =bleu_data[pair]["""src"""] A : Optional[Any] =bleu_data[pair]["""tgt"""] A : Dict =tokenizer(__lowercase , return_tensors='pt' , truncation=__lowercase , padding='longest' ).to(__lowercase ) A : Optional[int] =model.generate( input_ids=batch.input_ids , num_beams=8 , ) A : Optional[Any] =tokenizer.batch_decode( __lowercase , skip_special_tokens=__lowercase , clean_up_tokenization_spaces=__lowercase ) A : Tuple =calculate_bleu(__lowercase , __lowercase ) print(__lowercase ) self.assertGreaterEqual(scores['bleu'] , __lowercase )
305
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() a : Dict = logging.get_logger(__name__) a : Tuple = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowerCamelCase__ ( __lowerCamelCase : Dict ): if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase : Union[str, Any] = k.replace(__lowerCamelCase , __lowerCamelCase ) if k.startswith("""encoder""" ): __UpperCAmelCase : List[str] = k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase : Optional[Any] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : Union[str, Any] = k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase : Optional[int] = k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase : List[Any] = k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase : Any = k.replace("""norm3""" , """final_layer_norm""" ) return k def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Optional[Any] = [ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase : Dict = sd.pop(__lowerCamelCase ) __UpperCAmelCase : List[str] = k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase : List[str] = v a : Optional[int] = ["START"] @torch.no_grad() def lowerCamelCase__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : str ): __UpperCAmelCase : str = torch.load(__lowerCamelCase , map_location="""cpu""" ) __UpperCAmelCase : Tuple = model["""model"""] __UpperCAmelCase : int = BlenderbotConfig.from_json_file(__lowerCamelCase ) __UpperCAmelCase : List[str] = BlenderbotForConditionalGeneration(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = m.model.state_dict().keys() __UpperCAmelCase : Any = [] __UpperCAmelCase : Any = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase : int = rename_state_dict_key(__lowerCamelCase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase : str = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__lowerCamelCase ) m.model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) m.half() m.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) a : Any = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
63
0
from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class __lowercase ( lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
637
def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ): __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : List[str] = len(__lowerCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Union[str, Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None __UpperCAmelCase : str = sorted_collection[point] if current_item == item: return point else: if point < left: __UpperCAmelCase : Optional[Any] = left __UpperCAmelCase : Tuple = point elif point > right: __UpperCAmelCase : Optional[Any] = right __UpperCAmelCase : Dict = point else: if item < current_item: __UpperCAmelCase : Union[str, Any] = point - 1 else: __UpperCAmelCase : str = point + 1 return None def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowerCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) elif point > right: return interpolation_search_by_recursion(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , point - 1 ) else: return interpolation_search_by_recursion( __lowerCamelCase , __lowerCamelCase , point + 1 , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : int ): if collection != sorted(__lowerCamelCase ): raise ValueError("""Collection must be ascending sorted""" ) return True if __name__ == "__main__": import sys a : Optional[Any] = 0 if debug == 1: a : Optional[Any] = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") a : Tuple = 67 a : List[Any] = interpolation_search(collection, target) if result is not None: print(f"""{target} found at positions: {result}""") else: print("Not found")
63
0
"""simple docstring""" def A_ ( __lowercase ): if num <= 0: raise ValueError('Input must be a positive integer' ) UpperCamelCase_ : int =[True] * (num + 1) UpperCamelCase_ : Tuple =2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCamelCase ): UpperCamelCase_ : str =False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE = int(input('Enter a positive integer: ').strip()) print(prime_sieve_eratosthenes(user_num))
357
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
63
0
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class A__(unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ) -> Any: debug_launcher(test_script.main ) def UpperCamelCase__ ( self ) -> Union[str, Any]: debug_launcher(test_ops.main )
540
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class a ( lowercase__ , lowercase__ ): """simple docstring""" a : Dict = 1 @register_to_config def __init__( self : int , __lowercase : int = 1000 , __lowercase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: # set `betas`, `alphas`, `timesteps` self.set_timesteps(__lowercase ) # standard deviation of the initial noise distribution __UpperCAmelCase : List[Any] = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __UpperCAmelCase : List[Any] = 4 # running values __UpperCAmelCase : str = [] def UpperCAmelCase ( self : Union[str, Any] , __lowercase : int , __lowercase : Union[str, torch.device] = None ) -> int: __UpperCAmelCase : int = num_inference_steps __UpperCAmelCase : Union[str, Any] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __UpperCAmelCase : Union[str, Any] = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __UpperCAmelCase : Dict = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __UpperCAmelCase : Dict = torch.sin(steps * math.pi / 2 ) ** 2 __UpperCAmelCase : List[Any] = (1.0 - self.betas**2) ** 0.5 __UpperCAmelCase : Tuple = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __UpperCAmelCase : Dict = timesteps.to(__lowercase ) __UpperCAmelCase : Optional[Any] = [] def UpperCAmelCase ( self : Optional[int] , __lowercase : torch.FloatTensor , __lowercase : int , __lowercase : torch.FloatTensor , __lowercase : bool = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __UpperCAmelCase : List[str] = (self.timesteps == timestep).nonzero().item() __UpperCAmelCase : Optional[Any] = timestep_index + 1 __UpperCAmelCase : List[str] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowercase ) if len(self.ets ) == 1: __UpperCAmelCase : Tuple = self.ets[-1] elif len(self.ets ) == 2: __UpperCAmelCase : Union[str, Any] = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __UpperCAmelCase : Union[str, Any] = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __UpperCAmelCase : List[Any] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __UpperCAmelCase : Union[str, Any] = self._get_prev_sample(__lowercase , __lowercase , __lowercase , __lowercase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowercase ) def UpperCAmelCase ( self : Optional[Any] , __lowercase : torch.FloatTensor , *__lowercase : Optional[Any] , **__lowercase : Any ) -> torch.FloatTensor: return sample def UpperCAmelCase ( self : Tuple , __lowercase : Tuple , __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Dict ) -> str: __UpperCAmelCase : int = self.alphas[timestep_index] __UpperCAmelCase : Tuple = self.betas[timestep_index] __UpperCAmelCase : Any = self.alphas[prev_timestep_index] __UpperCAmelCase : List[str] = self.betas[prev_timestep_index] __UpperCAmelCase : List[str] = (sample - sigma * ets) / max(__lowercase , 1e-8 ) __UpperCAmelCase : List[Any] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Tuple ) -> str: return self.config.num_train_timesteps
63
0
'''simple docstring''' def __snake_case ( lowerCamelCase_ : int = 10**12 ): '''simple docstring''' __magic_name__ = 1 __magic_name__ = 0 __magic_name__ = 1 __magic_name__ = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(F'''{solution() = }''')
664
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCamelCase__ ( ): __UpperCAmelCase : Union[str, Any] = ArgumentParser("""Transformers CLI tool""" , usage="""transformers-cli <command> [<args>]""" ) __UpperCAmelCase : Any = parser.add_subparsers(help="""transformers-cli command helpers""" ) # Register commands ConvertCommand.register_subcommand(__lowerCamelCase ) DownloadCommand.register_subcommand(__lowerCamelCase ) EnvironmentCommand.register_subcommand(__lowerCamelCase ) RunCommand.register_subcommand(__lowerCamelCase ) ServeCommand.register_subcommand(__lowerCamelCase ) UserCommands.register_subcommand(__lowerCamelCase ) AddNewModelCommand.register_subcommand(__lowerCamelCase ) AddNewModelLikeCommand.register_subcommand(__lowerCamelCase ) LfsCommands.register_subcommand(__lowerCamelCase ) PTtoTFCommand.register_subcommand(__lowerCamelCase ) # Let's go __UpperCAmelCase : Optional[Any] = parser.parse_args() if not hasattr(__lowerCamelCase , """func""" ): parser.print_help() exit(1 ) # Run __UpperCAmelCase : Tuple = args.func(__lowerCamelCase ) service.run() if __name__ == "__main__": main()
63
0
import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase : '''simple docstring''' def __init__(self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=128 , __a=32 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> List[str]: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_input_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_labels UpperCAmelCase__ = num_choices UpperCAmelCase__ = scope def UpperCamelCase__ (self ) -> Any: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_input_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def UpperCamelCase__ (self ) -> int: """simple docstring""" ( UpperCAmelCase__ ) = self.prepare_config_and_inputs() UpperCAmelCase__ = True UpperCAmelCase__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> Any: """simple docstring""" UpperCAmelCase__ = NezhaModel(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase ) UpperCAmelCase__ = model(__lowercase , token_type_ids=__lowercase ) UpperCAmelCase__ = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a , __a , __a , ) -> Tuple: """simple docstring""" UpperCAmelCase__ = True UpperCAmelCase__ = NezhaModel(__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , ) UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , encoder_hidden_states=__lowercase , ) UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> Dict: """simple docstring""" UpperCAmelCase__ = NezhaForMaskedLM(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> List[str]: """simple docstring""" UpperCAmelCase__ = NezhaForNextSentencePrediction(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> Any: """simple docstring""" UpperCAmelCase__ = NezhaForPreTraining(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase , next_sentence_label=__lowercase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> str: """simple docstring""" UpperCAmelCase__ = NezhaForQuestionAnswering(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , start_positions=__lowercase , end_positions=__lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> int: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = NezhaForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = NezhaForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ (self , __a , __a , __a , __a , __a , __a , __a ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = self.num_choices UpperCAmelCase__ = NezhaForMultipleChoice(config=__lowercase ) model.to(__lowercase ) model.eval() UpperCAmelCase__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase__ = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase__ (self ) -> int: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() ( UpperCAmelCase__ ) = config_and_inputs UpperCAmelCase__ = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowercase ( lowercase__ , lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { 'feature-extraction': NezhaModel, 'fill-mask': NezhaForMaskedLM, 'question-answering': NezhaForQuestionAnswering, 'text-classification': NezhaForSequenceClassification, 'token-classification': NezhaForTokenClassification, 'zero-shot': NezhaForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = True def UpperCamelCase__ (self , __a , __a , __a=False ) -> Dict: """simple docstring""" UpperCAmelCase__ = super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if return_labels: if model_class in get_values(__lowercase ): UpperCAmelCase__ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__lowercase ) UpperCAmelCase__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowercase ) return inputs_dict def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = NezhaModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__lowercase , hidden_size=37 ) def UpperCamelCase__ (self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__lowercase ) def UpperCamelCase__ (self ) -> int: """simple docstring""" ( UpperCAmelCase__ ) = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCAmelCase__ = None self.model_tester.create_and_check_model_as_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowercase ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__lowercase ) def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*__lowercase ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__lowercase ) def UpperCamelCase__ (self ) -> Any: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowercase ) def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowercase ) def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowercase ) @slow def UpperCamelCase__ (self ) -> Any: """simple docstring""" for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = NezhaModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @slow @require_torch_gpu def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return UpperCAmelCase__ = True UpperCAmelCase__ = model_class(config=__lowercase ) UpperCAmelCase__ = self._prepare_for_class(__lowercase , __lowercase ) UpperCAmelCase__ = torch.jit.trace( __lowercase , (inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowercase , os.path.join(__lowercase , 'bert.pt' ) ) UpperCAmelCase__ = torch.jit.load(os.path.join(__lowercase , 'bert.pt' ) , map_location=__lowercase ) loaded(inputs_dict['input_ids'].to(__lowercase ) , inputs_dict['attention_mask'].to(__lowercase ) ) @require_torch class lowercase ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ (self ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = NezhaModel.from_pretrained('sijunhe/nezha-cn-base' ) UpperCAmelCase__ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase )[0] UpperCAmelCase__ = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , __lowercase ) UpperCAmelCase__ = torch.tensor([[[0.06_85, 0.24_41, 0.11_02], [0.06_00, 0.19_06, 0.13_49], [0.02_21, 0.08_19, 0.05_86]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowercase , atol=1E-4 ) ) @slow def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = NezhaForMaskedLM.from_pretrained('sijunhe/nezha-cn-base' ) UpperCAmelCase__ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCAmelCase__ = model(__lowercase , attention_mask=__lowercase )[0] UpperCAmelCase__ = torch.Size((1, 6, 21128) ) self.assertEqual(output.shape , __lowercase ) UpperCAmelCase__ = torch.tensor( [[-2.79_39, -1.79_02, -2.21_89], [-2.85_85, -1.89_08, -2.37_23], [-2.64_99, -1.77_50, -2.25_58]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowercase , atol=1E-4 ) )
146
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : List[str] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Dict = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
63
0
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = 0 def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = AutoImageProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) self.assertIsInstance(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Any ): with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = Path(__lowercase ) / """preprocessor_config.json""" __lowercase = Path(__lowercase ) / """config.json""" json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} ,open(__lowercase ,'''w''' ) ,) json.dump({'''model_type''': '''clip'''} ,open(__lowercase ,'''w''' ) ) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Any ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = Path(__lowercase ) / """preprocessor_config.json""" __lowercase = Path(__lowercase ) / """config.json""" json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} ,open(__lowercase ,'''w''' ) ,) json.dump({'''model_type''': '''clip'''} ,open(__lowercase ,'''w''' ) ) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = CLIPConfig() # Create a dummy config file with image_proceesor_type __lowercase = Path(__lowercase ) / """preprocessor_config.json""" __lowercase = Path(__lowercase ) / """config.json""" json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} ,open(__lowercase ,'''w''' ) ,) json.dump({'''model_type''': '''clip'''} ,open(__lowercase ,'''w''' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __lowercase = AutoImageProcessor.from_pretrained(__lowercase ).to_dict() config_dict.pop('''image_processor_type''' ) __lowercase = CLIPImageProcessor(**__lowercase ) # save in new folder model_config.save_pretrained(__lowercase ) config.save_pretrained(__lowercase ) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ) # make sure private variable is not incorrectly saved __lowercase = json.loads(config.to_json_string() ) self.assertTrue('''_processor_class''' not in dict_as_saved ) self.assertIsInstance(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = Path(__lowercase ) / """preprocessor_config.json""" json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} ,open(__lowercase ,'''w''' ) ,) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): with self.assertRaisesRegex( __lowercase ,'''clip-base is not a local folder and is not a valid model identifier''' ): __lowercase = AutoImageProcessor.from_pretrained('''clip-base''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): with self.assertRaisesRegex( __lowercase ,r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): __lowercase = AutoImageProcessor.from_pretrained(__lowercase ,revision='''aaaaaa''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): with self.assertRaisesRegex( __lowercase ,'''hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.''' ,): __lowercase = AutoImageProcessor.from_pretrained('''hf-internal-testing/config-no-model''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__lowercase ): __lowercase = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(__lowercase ): __lowercase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' ,trust_remote_code=__lowercase ) __lowercase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' ,trust_remote_code=__lowercase ) self.assertEqual(image_processor.__class__.__name__ ,'''NewImageProcessor''' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__lowercase ) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ,trust_remote_code=__lowercase ) self.assertEqual(reloaded_image_processor.__class__.__name__ ,'''NewImageProcessor''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): try: AutoConfig.register('''custom''' ,__lowercase ) AutoImageProcessor.register(__lowercase ,__lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__lowercase ): AutoImageProcessor.register(__lowercase ,__lowercase ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = Path(__lowercase ) / """preprocessor_config.json""" __lowercase = Path(__lowercase ) / """config.json""" json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} ,open(__lowercase ,'''w''' ) ,) json.dump({'''model_type''': '''clip'''} ,open(__lowercase ,'''w''' ) ) __lowercase = CustomImageProcessor.from_pretrained(__lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__lowercase ) __lowercase = AutoImageProcessor.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): class lowercase_ (lowercase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = True try: AutoConfig.register('''custom''' ,__lowercase ) AutoImageProcessor.register(__lowercase ,__lowercase ) # If remote code is not set, the default is to use local __lowercase = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) self.assertEqual(image_processor.__class__.__name__ ,'''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. __lowercase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' ,trust_remote_code=__lowercase ) self.assertEqual(image_processor.__class__.__name__ ,'''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub __lowercase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' ,trust_remote_code=__lowercase ) self.assertEqual(image_processor.__class__.__name__ ,'''NewImageProcessor''' ) self.assertTrue(not hasattr(__lowercase ,'''is_local''' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
41
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys a : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
import importlib.metadata import operator import re import sys from typing import Optional from packaging import version SCREAMING_SNAKE_CASE :Optional[int] = { "<": operator.lt, "<=": operator.le, "==": operator.eq, "!=": operator.ne, ">=": operator.ge, ">": operator.gt, } def _lowerCAmelCase ( lowerCAmelCase_ :List[str] , lowerCAmelCase_ :List[str] , lowerCAmelCase_ :str , lowerCAmelCase_ :int , lowerCAmelCase_ :Optional[Any] , lowerCAmelCase_ :str )->Union[str, Any]: '''simple docstring''' if got_ver is None or want_ver is None: raise ValueError( F'''Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider''' F''' reinstalling {pkg}.''' ) if not ops[op](version.parse(__lowerCamelCase ) , version.parse(__lowerCamelCase ) ): raise ImportError( F'''{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}''' ) def _lowerCAmelCase ( lowerCAmelCase_ :str , lowerCAmelCase_ :Optional[str] = None )->Any: '''simple docstring''' snake_case_ = F'''\n{hint}''' if hint is not None else """""" # non-versioned check if re.match(r"^[\w_\-\d]+$" , __lowerCamelCase ): snake_case_ = requirement, None, None else: snake_case_ = re.findall(r"^([^!=<>\s]+)([\s!=<>]{1,2}.+)" , __lowerCamelCase ) if not match: raise ValueError( "requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but" F''' got {requirement}''' ) snake_case_ = match[0] snake_case_ = want_full.split("," ) # there could be multiple requirements snake_case_ = {} for w in want_range: snake_case_ = re.findall(r"^([\s!=<>]{1,2})(.+)" , __lowerCamelCase ) if not match: raise ValueError( "requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23," F''' but got {requirement}''' ) snake_case_ = match[0] snake_case_ = want_ver if op not in ops: raise ValueError(F'''{requirement}: need one of {list(ops.keys() )}, but got {op}''' ) # special case if pkg == "python": snake_case_ = """.""".join([str(__lowerCamelCase ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return # check if any version is installed try: snake_case_ = importlib.metadata.version(__lowerCamelCase ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( F'''The \'{requirement}\' distribution was not found and is required by this application. {hint}''' ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowerCAmelCase ( lowerCAmelCase_ :List[Any] )->Tuple: '''simple docstring''' snake_case_ = """Try: pip install transformers -U or pip install -e '.[dev]' if you're working with git main""" return require_version(__lowerCamelCase , __lowerCamelCase )
283
def lowerCamelCase__ ( __lowerCamelCase : int ): if num <= 0: raise ValueError("""Input must be a positive integer""" ) __UpperCAmelCase : int = [True] * (num + 1) __UpperCAmelCase : Tuple = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCamelCase ): __UpperCAmelCase : str = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() a : Any = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
63
0
from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class __a : __snake_case : int __snake_case : int class __a : def __init__( self : Optional[int] , UpperCAmelCase : int ): lowerCAmelCase_ : list[list[Edge]] = [[] for _ in range(__lowercase )] lowerCAmelCase_ : Optional[Any] = size def __getitem__( self : int , UpperCAmelCase : int ): return iter(self._graph[vertex] ) @property def A ( self : Optional[Any] ): return self._size def A ( self : List[str] , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : int ): if weight not in (0, 1): raise ValueError("""Edge weight must be either 0 or 1.""" ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("""Vertex indexes must be in [0; size).""" ) self._graph[from_vertex].append(Edge(__lowercase , __lowercase ) ) def A ( self : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : int ): lowerCAmelCase_ : List[str] = deque([start_vertex] ) lowerCAmelCase_ : list[int | None] = [None] * self.size lowerCAmelCase_ : Optional[int] = 0 while queue: lowerCAmelCase_ : Tuple = queue.popleft() lowerCAmelCase_ : Tuple = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: lowerCAmelCase_ : Any = current_distance + edge.weight lowerCAmelCase_ : Tuple = distances[edge.destination_vertex] if ( isinstance(__lowercase , __lowercase ) and new_distance >= dest_vertex_distance ): continue lowerCAmelCase_ : Optional[int] = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("""No path from start_vertex to finish_vertex.""" ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
600
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Union[str, Any] = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class a ( lowercase__ ): """simple docstring""" a : Optional[int] = 'git_vision_model' def __init__( self : str , __lowercase : List[str]=768 , __lowercase : List[str]=3072 , __lowercase : List[Any]=12 , __lowercase : Dict=12 , __lowercase : int=3 , __lowercase : Any=224 , __lowercase : Optional[int]=16 , __lowercase : Dict="quick_gelu" , __lowercase : Any=1e-5 , __lowercase : str=0.0 , __lowercase : int=0.02 , **__lowercase : int , ) -> List[str]: super().__init__(**__lowercase ) __UpperCAmelCase : int = hidden_size __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Dict = num_attention_heads __UpperCAmelCase : int = num_channels __UpperCAmelCase : str = patch_size __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : int = initializer_range __UpperCAmelCase : Tuple = attention_dropout __UpperCAmelCase : Optional[int] = layer_norm_eps __UpperCAmelCase : List[Any] = hidden_act @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : Union[str, os.PathLike] , **__lowercase : List[str] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__lowercase ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("""model_type""" ) == "git": __UpperCAmelCase : str = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class a ( lowercase__ ): """simple docstring""" a : List[str] = 'git' def __init__( self : Optional[int] , __lowercase : List[Any]=None , __lowercase : Tuple=30522 , __lowercase : str=768 , __lowercase : Optional[int]=6 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[int]=3072 , __lowercase : List[str]="gelu" , __lowercase : Tuple=0.1 , __lowercase : Optional[int]=0.1 , __lowercase : Optional[int]=1024 , __lowercase : Union[str, Any]=0.02 , __lowercase : Optional[Any]=1e-1_2 , __lowercase : List[Any]=0 , __lowercase : Dict="absolute" , __lowercase : Dict=True , __lowercase : Any=False , __lowercase : Optional[int]=101 , __lowercase : str=102 , __lowercase : Union[str, Any]=None , **__lowercase : Dict , ) -> Tuple: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: __UpperCAmelCase : Optional[int] = {} logger.info("""vision_config is None. initializing the GitVisionConfig with default values.""" ) __UpperCAmelCase : Tuple = GitVisionConfig(**__lowercase ) __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Union[str, Any] = hidden_size __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : List[str] = hidden_act __UpperCAmelCase : str = intermediate_size __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : int = max_position_embeddings __UpperCAmelCase : str = initializer_range __UpperCAmelCase : str = layer_norm_eps __UpperCAmelCase : Union[str, Any] = position_embedding_type __UpperCAmelCase : Dict = use_cache __UpperCAmelCase : int = tie_word_embeddings __UpperCAmelCase : Optional[int] = num_image_with_embedding __UpperCAmelCase : Optional[int] = bos_token_id __UpperCAmelCase : List[Any] = eos_token_id def UpperCAmelCase ( self : str ) -> int: __UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : List[str] = self.vision_config.to_dict() __UpperCAmelCase : Union[str, Any] = self.__class__.model_type return output
63
0
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _lowerCamelCase = random.Random() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any , UpperCamelCase__: List[Any]=1.0 , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: int=None ): if rng is None: SCREAMING_SNAKE_CASE__ = global_rng SCREAMING_SNAKE_CASE__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class UpperCamelCase_ ( unittest.TestCase ): def __init__( self :str , __A :List[Any] , __A :Dict=7 , __A :Tuple=400 , __A :Optional[int]=2000 , __A :List[Any]=1 , __A :List[str]=0.0 , __A :str=1_6000 , __A :Tuple=True , __A :List[str]=True , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = min_seq_length SCREAMING_SNAKE_CASE__ = max_seq_length SCREAMING_SNAKE_CASE__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) SCREAMING_SNAKE_CASE__ = feature_size SCREAMING_SNAKE_CASE__ = padding_value SCREAMING_SNAKE_CASE__ = sampling_rate SCREAMING_SNAKE_CASE__ = return_attention_mask SCREAMING_SNAKE_CASE__ = do_normalize def _snake_case ( self :Optional[int] ) -> Dict: """simple docstring""" return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _snake_case ( self :Optional[int] , __A :int=False , __A :str=False ) -> int: """simple docstring""" def _flatten(__A :Optional[Any] ): return list(itertools.chain(*__lowercase ) ) if equal_length: SCREAMING_SNAKE_CASE__ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size SCREAMING_SNAKE_CASE__ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: SCREAMING_SNAKE_CASE__ = [np.asarray(__lowercase ) for x in speech_inputs] return speech_inputs class UpperCamelCase_ ( lowercase__ , unittest.TestCase ): lowerCamelCase_ = WavaVecaFeatureExtractor def _snake_case ( self :List[Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractionTester(self ) def _snake_case ( self :Optional[int] , __A :Any ) -> List[str]: """simple docstring""" self.assertTrue(np.all(np.mean(__lowercase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__lowercase , axis=0 ) - 1 ) < 1E-3 ) ) def _snake_case ( self :Dict ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE__ = [np.asarray(__lowercase ) for speech_input in speech_inputs] # Test not batched input SCREAMING_SNAKE_CASE__ = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values SCREAMING_SNAKE_CASE__ = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(__lowercase , __lowercase , atol=1E-3 ) ) # Test batched SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , return_tensors="""np""" ).input_values SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(__lowercase , __lowercase ): self.assertTrue(np.allclose(__lowercase , __lowercase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in (800, 800, 800)] SCREAMING_SNAKE_CASE__ = np.asarray(__lowercase ) SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , return_tensors="""np""" ).input_values SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(__lowercase , __lowercase ): self.assertTrue(np.allclose(__lowercase , __lowercase , atol=1E-3 ) ) def _snake_case ( self :Tuple ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE__ = ["""longest""", """max_length""", """do_not_pad"""] SCREAMING_SNAKE_CASE__ = [None, 1600, None] for max_length, padding in zip(__lowercase , __lowercase ): SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , padding=__lowercase , max_length=__lowercase , return_tensors="""np""" ) SCREAMING_SNAKE_CASE__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE__ = range(800 , 1400 , 200 ) SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in lengths] SCREAMING_SNAKE_CASE__ = ["""longest""", """max_length""", """do_not_pad"""] SCREAMING_SNAKE_CASE__ = [None, 1600, None] for max_length, padding in zip(__lowercase , __lowercase ): SCREAMING_SNAKE_CASE__ = feat_extract(__lowercase , max_length=__lowercase , padding=__lowercase ) SCREAMING_SNAKE_CASE__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _snake_case ( self :str ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE__ = feat_extract( __lowercase , truncation=__lowercase , max_length=1000 , padding="""max_length""" , return_tensors="""np""" ) SCREAMING_SNAKE_CASE__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def _snake_case ( self :Dict ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE__ = feat_extract( __lowercase , truncation=__lowercase , max_length=1000 , padding="""longest""" , return_tensors="""np""" ) SCREAMING_SNAKE_CASE__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) SCREAMING_SNAKE_CASE__ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE__ = feat_extract( __lowercase , truncation=__lowercase , max_length=2000 , padding="""longest""" , return_tensors="""np""" ) SCREAMING_SNAKE_CASE__ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def _snake_case ( self :str ) -> int: """simple docstring""" import torch SCREAMING_SNAKE_CASE__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE__ = np.random.rand(100 ).astype(np.floataa ) SCREAMING_SNAKE_CASE__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: SCREAMING_SNAKE_CASE__ = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) SCREAMING_SNAKE_CASE__ = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def _snake_case ( self :Tuple ) -> Dict: """simple docstring""" for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: SCREAMING_SNAKE_CASE__ = WavaVecaConfig.from_pretrained(__lowercase ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(__lowercase ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == """layer""" )
6
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = BarthezTokenizer a : Any = BarthezTokenizerFast a : Union[str, Any] = True a : Union[str, Any] = True def UpperCAmelCase ( self : Dict ) -> Any: super().setUp() __UpperCAmelCase : Optional[int] = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=__lowercase ) __UpperCAmelCase : str = tokenizer def UpperCAmelCase ( self : Optional[int] ) -> Tuple: __UpperCAmelCase : Dict = """<pad>""" __UpperCAmelCase : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> str: __UpperCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__lowercase ) , 101122 ) def UpperCAmelCase ( self : Any ) -> List[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __UpperCAmelCase : str = [0, 57, 3018, 70307, 91, 2] __UpperCAmelCase : List[Any] = self.tokenizer( __lowercase , max_length=len(__lowercase ) , padding=__lowercase , truncation=__lowercase , return_tensors="""pt""" ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __UpperCAmelCase : int = batch.input_ids.tolist()[0] self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> Tuple: if not self.test_rust_tokenizer: return __UpperCAmelCase : Union[str, Any] = self.get_tokenizer() __UpperCAmelCase : Optional[Any] = self.get_rust_tokenizer() __UpperCAmelCase : int = """I was born in 92000, and this is falsé.""" __UpperCAmelCase : Union[str, Any] = tokenizer.tokenize(__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : List[Any] = rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __UpperCAmelCase : Dict = self.get_rust_tokenizer() __UpperCAmelCase : str = tokenizer.encode(__lowercase ) __UpperCAmelCase : Tuple = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: # fmt: off __UpperCAmelCase : str = {"""input_ids""": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __UpperCAmelCase : int = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=__lowercase , )
63
0
import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters lowercase : Dict = False lowercase : str = False def lowerCAmelCase__ ( _a : Namespace ): return TrainCommand(__lowerCamelCase ) class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' @staticmethod def _lowerCAmelCase ( _SCREAMING_SNAKE_CASE ) -> Dict: snake_case_ : Optional[int] = parser.add_parser("train" , help="CLI tool to train a model on a task." ) train_parser.add_argument( "--train_data" , type=__lowercase , required=__lowercase , help="path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences." , ) train_parser.add_argument( "--column_label" , type=__lowercase , default=0 , help="Column of the dataset csv file with example labels." ) train_parser.add_argument( "--column_text" , type=__lowercase , default=1 , help="Column of the dataset csv file with example texts." ) train_parser.add_argument( "--column_id" , type=__lowercase , default=2 , help="Column of the dataset csv file with example ids." ) train_parser.add_argument( "--skip_first_row" , action="store_true" , help="Skip the first row of the csv file (headers)." ) train_parser.add_argument("--validation_data" , type=__lowercase , default="" , help="path to validation dataset." ) train_parser.add_argument( "--validation_split" , type=__lowercase , default=0.1 , help="if validation dataset is not provided, fraction of train dataset to use as validation dataset." , ) train_parser.add_argument("--output" , type=__lowercase , default="./" , help="path to saved the trained model." ) train_parser.add_argument( "--task" , type=__lowercase , default="text_classification" , help="Task to train the model on." ) train_parser.add_argument( "--model" , type=__lowercase , default="bert-base-uncased" , help="Model's name or path to stored model." ) train_parser.add_argument("--train_batch_size" , type=__lowercase , default=32 , help="Batch size for training." ) train_parser.add_argument("--valid_batch_size" , type=__lowercase , default=64 , help="Batch size for validation." ) train_parser.add_argument("--learning_rate" , type=__lowercase , default=3e-5 , help="Learning rate." ) train_parser.add_argument("--adam_epsilon" , type=__lowercase , default=1e-08 , help="Epsilon for Adam optimizer." ) train_parser.set_defaults(func=__lowercase ) def __init__( self , _SCREAMING_SNAKE_CASE ) -> Dict: snake_case_ : Union[str, Any] = logging.get_logger("transformers-cli/training" ) snake_case_ : Any = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=__lowercase ) snake_case_ : List[Any] = args.output snake_case_ : List[Any] = args.column_label snake_case_ : int = args.column_text snake_case_ : Union[str, Any] = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": snake_case_ : Any = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) snake_case_ : Any = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) snake_case_ : int = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) snake_case_ : List[str] = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) snake_case_ : Optional[Any] = args.validation_split snake_case_ : Dict = args.train_batch_size snake_case_ : List[Any] = args.valid_batch_size snake_case_ : List[str] = args.learning_rate snake_case_ : Any = args.adam_epsilon def _lowerCAmelCase ( self ) -> List[Any]: if self.framework == "tf": return self.run_tf() return self.run_torch() def _lowerCAmelCase ( self ) -> Optional[Any]: raise NotImplementedError def _lowerCAmelCase ( self ) -> Optional[int]: self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
568
from __future__ import annotations import math def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : bool , __lowerCamelCase : list[int] , __lowerCamelCase : float ): if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(__lowerCamelCase ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , ) def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = [90, 23, 6, 33, 21, 65, 123, 34423] __UpperCAmelCase : str = math.log(len(__lowerCamelCase ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
import argparse from collections import defaultdict import yaml _lowercase : Any ="docs/source/en/_toctree.yml" def A__ ( lowercase: Union[str, Any] ) -> Optional[Any]: A : List[str] =defaultdict(__lowerCamelCase ) A : List[str] =[] A : Optional[int] =[] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({'local': doc['local'], 'title': doc['title']} ) else: new_doc_list.append(__lowerCamelCase ) A : Dict =new_doc_list A : Optional[int] =[key for key, value in counts.items() if value > 1] A : Dict =[] for duplicate_key in duplicates: A : Optional[Any] =list({doc['title'] for doc in doc_list if doc['local'] == duplicate_key} ) if len(__lowerCamelCase ) > 1: raise ValueError( F'{duplicate_key} is present several times in the documentation table of content at ' '`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ' 'others.' ) # Only add this once new_doc.append({'local': duplicate_key, 'title': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if 'local' not in counts or counts[doc['local']] == 1] ) A : int =sorted(__lowerCamelCase, key=lambda lowercase : s["title"].lower() ) # "overview" gets special treatment and is always first if len(__lowerCamelCase ) > 1: raise ValueError('{doc_list} has two \'overview\' docs which is not allowed.' ) overview_doc.extend(__lowerCamelCase ) # Sort return overview_doc def A__ ( lowercase: Optional[Any]=False ) -> Tuple: with open(__lowerCamelCase, encoding='utf-8' ) as f: A : Tuple =yaml.safe_load(f.read() ) # Get to the API doc A : List[str] =0 while content[api_idx]["title"] != "API": api_idx += 1 A : Any =content[api_idx]["""sections"""] # Then to the model doc A : Optional[int] =0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 A : str =api_doc[scheduler_idx]["""sections"""] A : Dict =clean_doc_toc(__lowerCamelCase ) A : List[str] =False if new_scheduler_doc != scheduler_doc: A : Optional[int] =True if overwrite: A : int =new_scheduler_doc if diff: if overwrite: A : str =api_doc with open(__lowerCamelCase, 'w', encoding='utf-8' ) as f: f.write(yaml.dump(__lowerCamelCase, allow_unicode=__lowerCamelCase ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) def A__ ( lowercase: List[Any]=False ) -> str: with open(__lowerCamelCase, encoding='utf-8' ) as f: A : Union[str, Any] =yaml.safe_load(f.read() ) # Get to the API doc A : Optional[int] =0 while content[api_idx]["title"] != "API": api_idx += 1 A : Tuple =content[api_idx]["""sections"""] # Then to the model doc A : Optional[Any] =0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 A : Tuple =False A : int =api_doc[pipeline_idx]["""sections"""] A : Optional[Any] =[] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: A : Optional[Any] =pipeline_doc["""section"""] A : List[Any] =clean_doc_toc(__lowerCamelCase ) if overwrite: A : List[str] =new_sub_pipeline_doc new_pipeline_docs.append(__lowerCamelCase ) # sort overall pipeline doc A : Any =clean_doc_toc(__lowerCamelCase ) if new_pipeline_docs != pipeline_docs: A : Tuple =True if overwrite: A : Any =new_pipeline_docs if diff: if overwrite: A : List[Any] =api_doc with open(__lowerCamelCase, 'w', encoding='utf-8' ) as f: f.write(yaml.dump(__lowerCamelCase, allow_unicode=__lowerCamelCase ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) if __name__ == "__main__": _lowercase : Union[str, Any] =argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowercase : Tuple =parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
305
from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class a ( lowercase__ ): """simple docstring""" a : Optional[Any] = 'openai-gpt' a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __lowercase : Tuple=40478 , __lowercase : Tuple=512 , __lowercase : int=768 , __lowercase : Dict=12 , __lowercase : Union[str, Any]=12 , __lowercase : Optional[Any]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Optional[Any]=0.1 , __lowercase : Tuple=0.1 , __lowercase : Dict=1e-5 , __lowercase : Any=0.02 , __lowercase : List[str]="cls_index" , __lowercase : str=True , __lowercase : Dict=None , __lowercase : str=True , __lowercase : List[str]=0.1 , **__lowercase : List[Any] , ) -> List[Any]: __UpperCAmelCase : Optional[int] = vocab_size __UpperCAmelCase : Optional[Any] = n_positions __UpperCAmelCase : Optional[int] = n_embd __UpperCAmelCase : str = n_layer __UpperCAmelCase : Any = n_head __UpperCAmelCase : Tuple = afn __UpperCAmelCase : Any = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : str = attn_pdrop __UpperCAmelCase : str = layer_norm_epsilon __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : Optional[int] = summary_type __UpperCAmelCase : Optional[Any] = summary_use_proj __UpperCAmelCase : List[Any] = summary_activation __UpperCAmelCase : Union[str, Any] = summary_first_dropout __UpperCAmelCase : Dict = summary_proj_to_labels super().__init__(**__lowercase )
63
0
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": __a : int = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=512, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def _SCREAMING_SNAKE_CASE ( __lowercase : Any ) -> Any: """simple docstring""" if string == "True": return True elif string == "False": return False else: raise ValueError(f"could not parse string as bool {string}" ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) __a : List[str] = parser.parse_args() __a : Optional[Any] = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
637
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = KandinskyVaaInpaintPipeline a : Any = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] a : Any = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] a : Any = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] a : List[Any] = False @property def UpperCAmelCase ( self : int ) -> Dict: return 32 @property def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: return 32 @property def UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: return self.time_input_dim @property def UpperCAmelCase ( self : str ) -> List[str]: return self.time_input_dim * 4 @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return 100 @property def UpperCAmelCase ( self : Dict ) -> Any: torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } __UpperCAmelCase : int = UNetaDConditionModel(**__lowercase ) return model @property def UpperCAmelCase ( self : int ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self : Any ) -> List[Any]: __UpperCAmelCase : List[str] = self.dummy_unet __UpperCAmelCase : List[str] = self.dummy_movq __UpperCAmelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=__lowercase , set_alpha_to_one=__lowercase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowercase , ) __UpperCAmelCase : str = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def UpperCAmelCase ( self : str , __lowercase : Tuple , __lowercase : List[str]=0 ) -> Optional[Any]: __UpperCAmelCase : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __lowercase ) # create init_image __UpperCAmelCase : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowercase ) ).to(__lowercase ) __UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(__lowercase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask __UpperCAmelCase : Union[str, Any] = np.ones((64, 64) , dtype=np.floataa ) __UpperCAmelCase : List[str] = 0 if str(__lowercase ).startswith("""mps""" ): __UpperCAmelCase : List[str] = torch.manual_seed(__lowercase ) else: __UpperCAmelCase : Optional[int] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __UpperCAmelCase : Optional[Any] = { """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = """cpu""" __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**__lowercase ) __UpperCAmelCase : Tuple = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(__lowercase ) ) __UpperCAmelCase : Tuple = output.images __UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(__lowercase ) , return_dict=__lowercase , )[0] __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] __UpperCAmelCase : str = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __UpperCAmelCase : Optional[Any] = np.array( [0.50_775_903, 0.49_527_195, 0.48_824_543, 0.50_192_237, 0.48_644_906, 0.49_373_814, 0.4_780_598, 0.47_234_827, 0.48_327_848] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def UpperCAmelCase ( self : str ) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy""" ) __UpperCAmelCase : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) __UpperCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa ) __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = """a hat""" __UpperCAmelCase : str = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowercase ) __UpperCAmelCase : Any = KandinskyVaaInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder-inpaint""" , torch_dtype=torch.floataa ) __UpperCAmelCase : int = pipeline.to(__lowercase ) pipeline.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = pipe_prior( __lowercase , generator=__lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() __UpperCAmelCase : Optional[int] = pipeline( image=__lowercase , mask_image=__lowercase , image_embeds=__lowercase , negative_image_embeds=__lowercase , generator=__lowercase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) __UpperCAmelCase : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase )
63
0
"""simple docstring""" import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand __SCREAMING_SNAKE_CASE = ( "4S 3H 2C 7S 5H", "9D 8H 2C 6S 7H", "2D 6D 9D TH 7D", "TC 8C 2S JH 6C", "JH 8S TH AH QH", "TS KS 5S 9S AC", "KD 6S 9D TH AD", "KS 8D 4D 9S 4S", # pair "8C 4S KH JS 4D", # pair "QH 8H KD JH 8S", # pair "KC 4H KS 2H 8D", # pair "KD 4S KC 3H 8S", # pair "AH 8S AS KC JH", # pair "3H 4C 4H 3S 2H", # 2 pairs "5S 5D 2C KH KH", # 2 pairs "3C KH 5D 5S KH", # 2 pairs "AS 3C KH AD KH", # 2 pairs "7C 7S 3S 7H 5S", # 3 of a kind "7C 7S KH 2H 7H", # 3 of a kind "AC KH QH AH AS", # 3 of a kind "2H 4D 3C AS 5S", # straight (low ace) "3C 5C 4C 2C 6H", # straight "6S 8S 7S 5H 9H", # straight "JS QS 9H TS KH", # straight "QC KH TS JS AH", # straight (high ace) "8C 9C 5C 3C TC", # flush "3S 8S 9S 5S KS", # flush "4C 5C 9C 8C KC", # flush "JH 8H AH KH QH", # flush "3D 2H 3H 2C 2D", # full house "2H 2C 3S 3H 3D", # full house "KH KC 3S 3H 3D", # full house "JC 6H JS JD JH", # 4 of a kind "JC 7H JS JD JH", # 4 of a kind "JC KH JS JD JH", # 4 of a kind "2S AS 4S 5S 3S", # straight flush (low ace) "2D 6D 3D 4D 5D", # straight flush "5C 6C 3C 7C 4C", # straight flush "JH 9H TH KH QH", # straight flush "JH AH TH KH QH", # royal flush (high ace straight flush) ) __SCREAMING_SNAKE_CASE = ( ("2H 3H 4H 5H 6H", "KS AS TS QS JS", "Loss"), ("2H 3H 4H 5H 6H", "AS AD AC AH JD", "Win"), ("AS AH 2H AD AC", "JS JD JC JH 3D", "Win"), ("2S AH 2H AS AC", "JS JD JC JH AD", "Loss"), ("2S AH 2H AS AC", "2H 3H 5H 6H 7H", "Win"), ("AS 3S 4S 8S 2S", "2H 3H 5H 6H 7H", "Win"), ("2H 3H 5H 6H 7H", "2S 3H 4H 5S 6C", "Win"), ("2S 3H 4H 5S 6C", "3D 4C 5H 6H 2S", "Tie"), ("2S 3H 4H 5S 6C", "AH AC 5H 6H AS", "Win"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H AS", "Loss"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H 7S", "Win"), ("6S AD 7H 4S AS", "AH AC 5H 6H 7S", "Loss"), ("2S AH 4H 5S KC", "AH AC 5H 6H 7S", "Loss"), ("2S 3H 6H 7S 9C", "7H 3C TH 6H 9S", "Loss"), ("4S 5H 6H TS AC", "3S 5H 6H TS AC", "Win"), ("2S AH 4H 5S 6C", "AD 4C 5H 6H 2C", "Tie"), ("AS AH 3H AD AC", "AS AH 2H AD AC", "Win"), ("AH AC 5H 5C QS", "AH AC 5H 5C KS", "Loss"), ("AH AC 5H 5C QS", "KH KC 5H 5C QS", "Win"), ("7C 7S KH 2H 7H", "3C 3S AH 2H 3H", "Win"), ("3C 3S AH 2H 3H", "7C 7S KH 2H 7H", "Loss"), ("6H 5H 4H 3H 2H", "5H 4H 3H 2H AH", "Win"), ("5H 4H 3H 2H AH", "5H 4H 3H 2H AH", "Tie"), ("5H 4H 3H 2H AH", "6H 5H 4H 3H 2H", "Loss"), ("AH AD KS KC AC", "AH KD KH AC KC", "Win"), ("2H 4D 3C AS 5S", "2H 4D 3C 6S 5S", "Loss"), ("2H 3S 3C 3H 2S", "3S 3C 2S 2H 2D", "Win"), ("4D 6D 5D 2D JH", "3S 8S 3H TC KH", "Loss"), ("4S 6C 8S 3S 7S", "AD KS 2D 7D 7C", "Loss"), ("6S 4C 7H 8C 3H", "5H JC AH 9D 9C", "Loss"), ("9D 9H JH TC QH", "3C 2S JS 5C 7H", "Win"), ("2H TC 8S AD 9S", "4H TS 7H 2C 5C", "Win"), ("9D 3S 2C 7S 7C", "JC TD 3C TC 9H", "Loss"), ) __SCREAMING_SNAKE_CASE = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", True), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", False), ("AS 3S 4S 8S 2S", True), ) __SCREAMING_SNAKE_CASE = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", False), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", True), ) __SCREAMING_SNAKE_CASE = ( ("2H 4D 3C AS 5S", True, [5, 4, 3, 2, 14]), ("2H 5D 3C AS 5S", False, [14, 5, 5, 3, 2]), ("JH QD KC AS TS", False, [14, 13, 12, 11, 10]), ("9D 3S 2C 7S 7C", False, [9, 7, 7, 3, 2]), ) __SCREAMING_SNAKE_CASE = ( ("JH AH TH KH QH", 0), ("JH 9H TH KH QH", 0), ("JC KH JS JD JH", 7), ("KH KC 3S 3H 3D", 6), ("8C 9C 5C 3C TC", 0), ("JS QS 9H TS KH", 0), ("7C 7S KH 2H 7H", 3), ("3C KH 5D 5S KH", 2), ("QH 8H KD JH 8S", 1), ("2D 6D 9D TH 7D", 0), ) __SCREAMING_SNAKE_CASE = ( ("JH AH TH KH QH", 23), ("JH 9H TH KH QH", 22), ("JC KH JS JD JH", 21), ("KH KC 3S 3H 3D", 20), ("8C 9C 5C 3C TC", 19), ("JS QS 9H TS KH", 18), ("7C 7S KH 2H 7H", 17), ("3C KH 5D 5S KH", 16), ("QH 8H KD JH 8S", 15), ("2D 6D 9D TH 7D", 14), ) def A_ ( ): UpperCamelCase_ : Optional[int] =randrange(len(__lowerCamelCase ) ), randrange(len(__lowerCamelCase ) ) UpperCamelCase_ : str =["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] UpperCamelCase_ : int =SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def A_ ( __lowercase = 1_00 ): return (generate_random_hand() for _ in range(__lowerCamelCase )) @pytest.mark.parametrize('hand, expected' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase )._is_flush() == expected @pytest.mark.parametrize('hand, expected' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase )._is_straight() == expected @pytest.mark.parametrize('hand, expected, card_values' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase , __lowercase ): UpperCamelCase_ : int =PokerHand(__lowerCamelCase ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('hand, expected' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase )._is_same_kind() == expected @pytest.mark.parametrize('hand, expected' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase )._hand_type == expected @pytest.mark.parametrize('hand, other, expected' , __lowerCamelCase ) def A_ ( __lowercase , __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase ).compare_with(PokerHand(__lowerCamelCase ) ) == expected @pytest.mark.parametrize('hand, other, expected' , generate_random_hands() ) def A_ ( __lowercase , __lowercase , __lowercase ): assert PokerHand(__lowerCamelCase ).compare_with(PokerHand(__lowerCamelCase ) ) == expected def A_ ( ): UpperCamelCase_ : Optional[Any] =[PokerHand(__lowerCamelCase ) for hand in SORTED_HANDS] UpperCamelCase_ : Any =poker_hands.copy() shuffle(__lowerCamelCase ) UpperCamelCase_ : str =chain(sorted(__lowerCamelCase ) ) for index, hand in enumerate(__lowerCamelCase ): assert hand == poker_hands[index] def A_ ( ): # Test that five high straights are compared correctly. UpperCamelCase_ : Any =[PokerHand('2D AC 3H 4H 5S' ), PokerHand('2S 3H 4H 5S 6C' )] pokerhands.sort(reverse=__lowerCamelCase ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def A_ ( ): # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. UpperCamelCase_ : List[Any] =PokerHand('2C 4S AS 3D 5C' ) UpperCamelCase_ : str =True UpperCamelCase_ : Dict =[5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def A_ ( ): # Problem number 54 from Project Euler # Testing from poker_hands.txt file UpperCamelCase_ : Tuple =0 UpperCamelCase_ : Tuple =os.path.abspath(os.path.dirname(__lowerCamelCase ) ) UpperCamelCase_ : Dict =os.path.join(__lowerCamelCase , 'poker_hands.txt' ) with open(__lowerCamelCase ) as file_hand: for line in file_hand: UpperCamelCase_ : Optional[Any] =line[:14].strip() UpperCamelCase_ : List[Any] =line[15:].strip() UpperCamelCase_ : List[str] =PokerHand(__lowerCamelCase ), PokerHand(__lowerCamelCase ) UpperCamelCase_ : Dict =player.compare_with(__lowerCamelCase ) if output == "Win": answer += 1 assert answer == 3_76
357
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a : List[Any] = True except ImportError: a : str = False try: from torch.hub import _get_torch_home a : List[Any] = _get_torch_home() except ImportError: a : int = os.path.expanduser( os.getenv("TORCH_HOME", os.path.join(os.getenv("XDG_CACHE_HOME", "~/.cache"), "torch")) ) a : Optional[Any] = os.path.join(torch_cache_home, "transformers") a : Optional[Any] = "https://cdn.huggingface.co" a : List[str] = "https://s3.amazonaws.com/models.huggingface.co/bert" a : Any = "/".join(str(Path(__file__).resolve()).split("/")[:-1]) a : Optional[int] = os.path.join(PATH, "config.yaml") a : Dict = os.path.join(PATH, "attributes.txt") a : Tuple = os.path.join(PATH, "objects.txt") a : Dict = os.getenv("PYTORCH_PRETRAINED_BERT_CACHE", default_cache_path) a : Dict = os.getenv("PYTORCH_TRANSFORMERS_CACHE", PYTORCH_PRETRAINED_BERT_CACHE) a : Optional[int] = os.getenv("TRANSFORMERS_CACHE", PYTORCH_TRANSFORMERS_CACHE) a : Any = "pytorch_model.bin" a : int = "config.yaml" def lowerCamelCase__ ( __lowerCamelCase : str=OBJECTS , __lowerCamelCase : Union[str, Any]=ATTRIBUTES ): __UpperCAmelCase : Union[str, Any] = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(""",""" )[0].lower().strip() ) __UpperCAmelCase : Dict = [] with open(__lowerCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(""",""" )[0].lower().strip() ) return vg_classes, vg_attrs def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : List[str] = OrderedDict() with open(__lowerCamelCase , """rb""" ) as f: __UpperCAmelCase : int = pkl.load(__lowerCamelCase )["""model"""] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : List[Any] = ckp.pop(__lowerCamelCase ) if isinstance(__lowerCamelCase , np.ndarray ): __UpperCAmelCase : Union[str, Any] = torch.tensor(__lowerCamelCase ) else: assert isinstance(__lowerCamelCase , torch.tensor ), type(__lowerCamelCase ) __UpperCAmelCase : List[str] = v return r class a : """simple docstring""" a : Dict = {} def __init__( self : Dict , __lowercase : dict , __lowercase : str = "root" , __lowercase : Any=0 ) -> Dict: __UpperCAmelCase : List[str] = name __UpperCAmelCase : str = level __UpperCAmelCase : int = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : List[str] = copy.deepcopy(__lowercase ) __UpperCAmelCase : Dict = copy.deepcopy(__lowercase ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : Union[str, Any] = Config(__lowercase , name=__lowercase , level=level + 1 ) __UpperCAmelCase : Union[str, Any] = v setattr(self , __lowercase , __lowercase ) __UpperCAmelCase : Any = d def __repr__( self : Optional[Any] ) -> Optional[int]: return str(list((self._pointer.keys()) ) ) def __setattr__( self : List[str] , __lowercase : List[str] , __lowercase : Tuple ) -> int: __UpperCAmelCase : int = val __UpperCAmelCase : List[str] = val __UpperCAmelCase : Union[str, Any] = key.split(""".""" ) __UpperCAmelCase : List[Any] = len(__lowercase ) - 1 __UpperCAmelCase : List[Any] = self._pointer if len(__lowercase ) > 1: for i, l in enumerate(__lowercase ): if hasattr(self , __lowercase ) and isinstance(getattr(self , __lowercase ) , __lowercase ): setattr(getattr(self , __lowercase ) , """.""".join(levels[i:] ) , __lowercase ) if l == last_level: __UpperCAmelCase : Union[str, Any] = val else: __UpperCAmelCase : Union[str, Any] = pointer[l] def UpperCAmelCase ( self : Tuple ) -> Optional[int]: return self._pointer def UpperCAmelCase ( self : str , __lowercase : Optional[int] , __lowercase : Any ) -> Optional[int]: with open(f"""{file_name}""" , """w""" ) as stream: dump(__lowercase , __lowercase ) def UpperCAmelCase ( self : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] ) -> Any: with open(f"""{file_name}""" , """w""" ) as stream: json.dump(__lowercase , __lowercase ) @staticmethod def UpperCAmelCase ( __lowercase : List[Any] ) -> Optional[Any]: with open(__lowercase ) as stream: __UpperCAmelCase : Any = load(__lowercase , Loader=__lowercase ) return data def __str__( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = """ """ if self._name != "root": __UpperCAmelCase : Optional[Any] = f"""{t * (self._level-1)}{self._name}:\n""" else: __UpperCAmelCase : List[Any] = """""" __UpperCAmelCase : Optional[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__lowercase , __lowercase ): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(__lowercase ).__name__})\n""" __UpperCAmelCase : int = level return r[:-1] @classmethod def UpperCAmelCase ( cls : List[str] , __lowercase : str , **__lowercase : Any ) -> Any: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) return cls(__lowercase ) @classmethod def UpperCAmelCase ( cls : Dict , __lowercase : str , **__lowercase : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase : int = kwargs.pop("""cache_dir""" , __lowercase ) __UpperCAmelCase : int = kwargs.pop("""force_download""" , __lowercase ) __UpperCAmelCase : str = kwargs.pop("""resume_download""" , __lowercase ) __UpperCAmelCase : Dict = kwargs.pop("""proxies""" , __lowercase ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""local_files_only""" , __lowercase ) if os.path.isdir(__lowercase ): __UpperCAmelCase : List[Any] = os.path.join(__lowercase , __lowercase ) elif os.path.isfile(__lowercase ) or is_remote_url(__lowercase ): __UpperCAmelCase : Tuple = pretrained_model_name_or_path else: __UpperCAmelCase : Optional[int] = hf_bucket_url(__lowercase , filename=__lowercase , use_cdn=__lowercase ) try: # Load from URL or cache if already cached __UpperCAmelCase : Optional[int] = cached_path( __lowercase , cache_dir=__lowercase , force_download=__lowercase , proxies=__lowercase , resume_download=__lowercase , local_files_only=__lowercase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : Optional[int] = Config.load_yaml(__lowercase ) except EnvironmentError: __UpperCAmelCase : str = """Can't load config for""" raise EnvironmentError(__lowercase ) if resolved_config_file == config_file: print("""loading configuration file from path""" ) else: print("""loading configuration file cache""" ) return Config.load_yaml(__lowercase ), kwargs def lowerCamelCase__ ( __lowerCamelCase : Dict ): __UpperCAmelCase : Optional[int] = torch.load("""dump.pt""" , map_location=in_tensor.device ) __UpperCAmelCase : Tuple = in_tensor.numpy() __UpperCAmelCase : Optional[int] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(__lowerCamelCase , __lowerCamelCase , rtol=0.0_1 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception("""tensors are all good""" ) # Hugging face functions below def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Tuple = urlparse(__lowerCamelCase ) return parsed.scheme in ("http", "https") def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : int=True ): __UpperCAmelCase : int = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : Optional[int] = """/""" not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=None , ): __UpperCAmelCase : Optional[int] = """python/{}""".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + "; ".join("""{}/{}""".format(__lowerCamelCase , __lowerCamelCase ) for k, v in user_agent.items() ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): ua += "; " + user_agent __UpperCAmelCase : List[str] = {"""user-agent""": ua} if resume_size > 0: __UpperCAmelCase : Union[str, Any] = """bytes=%d-""" % (resume_size,) __UpperCAmelCase : Union[str, Any] = requests.get(__lowerCamelCase , stream=__lowerCamelCase , proxies=__lowerCamelCase , headers=__lowerCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : List[str] = response.headers.get("""Content-Length""" ) __UpperCAmelCase : str = resume_size + int(__lowerCamelCase ) if content_length is not None else None __UpperCAmelCase : List[Any] = tqdm( unit="""B""" , unit_scale=__lowerCamelCase , total=__lowerCamelCase , initial=__lowerCamelCase , desc="""Downloading""" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(__lowerCamelCase ) ) temp_file.write(__lowerCamelCase ) progress.close() def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=10 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=None , __lowerCamelCase : List[str]=False , ): if cache_dir is None: __UpperCAmelCase : Optional[Any] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[str] = str(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) __UpperCAmelCase : List[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(__lowerCamelCase , allow_redirects=__lowerCamelCase , proxies=__lowerCamelCase , timeout=__lowerCamelCase ) if response.status_code == 200: __UpperCAmelCase : Dict = response.headers.get("""ETag""" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : List[str] = url_to_filename(__lowerCamelCase , __lowerCamelCase ) # get cache path to put the file __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , __lowerCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(__lowerCamelCase ): return cache_path else: __UpperCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(__lowerCamelCase ) , filename + """.*""" ) if not file.endswith(""".json""" ) and not file.endswith(""".lock""" ) ] if len(__lowerCamelCase ) > 0: return os.path.join(__lowerCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( """Cannot find the requested files in the cached path and outgoing traffic has been""" """ disabled. To enable model look-ups and downloads online, set 'local_files_only'""" """ to False.""" ) return None # From now on, etag is not None. if os.path.exists(__lowerCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : str = cache_path + """.lock""" with FileLock(__lowerCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(__lowerCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : int = cache_path + """.incomplete""" @contextmanager def _resumable_file_manager(): with open(__lowerCamelCase , """a+b""" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(__lowerCamelCase ): __UpperCAmelCase : List[Any] = os.stat(__lowerCamelCase ).st_size else: __UpperCAmelCase : List[Any] = 0 else: __UpperCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=__lowerCamelCase , delete=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( """%s not found in cache or force_download set to True, downloading to %s""" , __lowerCamelCase , temp_file.name , ) http_get( __lowerCamelCase , __lowerCamelCase , proxies=__lowerCamelCase , resume_size=__lowerCamelCase , user_agent=__lowerCamelCase , ) os.replace(temp_file.name , __lowerCamelCase ) __UpperCAmelCase : Any = {"""url""": url, """etag""": etag} __UpperCAmelCase : Union[str, Any] = cache_path + """.json""" with open(__lowerCamelCase , """w""" ) as meta_file: json.dump(__lowerCamelCase , __lowerCamelCase ) return cache_path def lowerCamelCase__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any]=None ): __UpperCAmelCase : Tuple = url.encode("""utf-8""" ) __UpperCAmelCase : Optional[Any] = shaaaa(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = url_hash.hexdigest() if etag: __UpperCAmelCase : int = etag.encode("""utf-8""" ) __UpperCAmelCase : List[str] = shaaaa(__lowerCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(""".h5""" ): filename += ".h5" return filename def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : int=None , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : str=False , ): if cache_dir is None: __UpperCAmelCase : List[str] = TRANSFORMERS_CACHE if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Any = str(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) if is_remote_url(__lowerCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : Tuple = get_from_cache( __lowerCamelCase , cache_dir=__lowerCamelCase , force_download=__lowerCamelCase , proxies=__lowerCamelCase , resume_download=__lowerCamelCase , user_agent=__lowerCamelCase , local_files_only=__lowerCamelCase , ) elif os.path.exists(__lowerCamelCase ): # File, and it exists. __UpperCAmelCase : Tuple = url_or_filename elif urlparse(__lowerCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("""file {} not found""".format(__lowerCamelCase ) ) else: # Something unknown raise ValueError("""unable to parse {} as a URL or as a local path""".format(__lowerCamelCase ) ) if extract_compressed_file: if not is_zipfile(__lowerCamelCase ) and not tarfile.is_tarfile(__lowerCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : int = os.path.split(__lowerCamelCase ) __UpperCAmelCase : Any = output_file.replace(""".""" , """-""" ) + """-extracted""" __UpperCAmelCase : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isdir(__lowerCamelCase ) and os.listdir(__lowerCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : str = output_path + """.lock""" with FileLock(__lowerCamelCase ): shutil.rmtree(__lowerCamelCase , ignore_errors=__lowerCamelCase ) os.makedirs(__lowerCamelCase ) if is_zipfile(__lowerCamelCase ): with ZipFile(__lowerCamelCase , """r""" ) as zip_file: zip_file.extractall(__lowerCamelCase ) zip_file.close() elif tarfile.is_tarfile(__lowerCamelCase ): __UpperCAmelCase : Any = tarfile.open(__lowerCamelCase ) tar_file.extractall(__lowerCamelCase ) tar_file.close() else: raise EnvironmentError("""Archive format of {} could not be identified""".format(__lowerCamelCase ) ) return output_path_extracted return output_path def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : int="," ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): with open(__lowerCamelCase ) as f: __UpperCAmelCase : List[Any] = eval(f.read() ) else: __UpperCAmelCase : List[str] = requests.get(__lowerCamelCase ) try: __UpperCAmelCase : int = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : str = eval(__lowerCamelCase ) except Exception: __UpperCAmelCase : List[Any] = data.split("""\n""" ) req.close() return data def lowerCamelCase__ ( __lowerCamelCase : Any ): __UpperCAmelCase : Optional[int] = requests.get(__lowerCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def lowerCamelCase__ ( __lowerCamelCase : str ): __UpperCAmelCase : int = url.split("""/""" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(__lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as stream: __UpperCAmelCase : List[str] = pkl.load(__lowerCamelCase ) __UpperCAmelCase : Dict = weights.pop("""model""" ) __UpperCAmelCase : Union[str, Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(__lowerCamelCase ) if "running_var" in k: __UpperCAmelCase : Optional[int] = torch.tensor([0] ) __UpperCAmelCase : Tuple = k.replace("""running_var""" , """num_batches_tracked""" ) __UpperCAmelCase : Any = zero return new def lowerCamelCase__ ( ): print(f"""{os.path.abspath(os.path.join(__lowerCamelCase , os.pardir ) )}/demo.ipynb""" ) def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]="RGB" ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ): __UpperCAmelCase : List[str] = cva.imread(__lowerCamelCase ) else: __UpperCAmelCase : int = get_image_from_url(__lowerCamelCase ) assert img is not None, f"""could not connect to: {im}""" __UpperCAmelCase : Any = cva.cvtColor(__lowerCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Optional[int] = img[:, :, ::-1] return img def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int=1 ): return (images[i : i + batch] for i in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ))
63
0
import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin __snake_case : Union[str, Any] = 1e-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class A__: """simple docstring""" def __init__( self , _lowercase , _lowercase=16 , _lowercase=13 , _lowercase=7 , _lowercase=14 , _lowercase=10 , _lowercase=19 , _lowercase=5 , _lowercase=4 , _lowercase=True , _lowercase=16 , _lowercase=2 , _lowercase=4 , _lowercase=4 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=[1, 2, 3, 4, 5] , _lowercase=25 , _lowercase=5 , ) -> Dict: a_ : Any = d_model a_ : Dict = parent a_ : List[Any] = batch_size a_ : List[str] = prediction_length a_ : Dict = context_length a_ : Any = cardinality a_ : Any = num_time_features a_ : Optional[Any] = lags_sequence a_ : Union[str, Any] = embedding_dimension a_ : List[str] = is_training a_ : Optional[int] = hidden_size a_ : List[Any] = num_hidden_layers a_ : List[str] = num_attention_heads a_ : Optional[Any] = intermediate_size a_ : List[str] = hidden_act a_ : Any = hidden_dropout_prob a_ : List[str] = attention_probs_dropout_prob a_ : Any = context_length a_ : int = prediction_length + label_length a_ : List[str] = label_length a_ : Union[str, Any] = moving_average a_ : Optional[int] = autocorrelation_factor def UpperCamelCase__ ( self ) -> List[Any]: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def UpperCamelCase__ ( self , _lowercase ) -> Optional[Any]: a_ : Tuple = config.context_length + max(config.lags_sequence ) a_ : str = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) a_ : str = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) a_ : Dict = floats_tensor([self.batch_size, _past_length] ) a_ : Dict = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs a_ : Optional[int] = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) a_ : Any = floats_tensor([self.batch_size, config.prediction_length] ) a_ : Union[str, Any] = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def UpperCamelCase__ ( self ) -> List[Any]: a_ : Any = self.get_config() a_ : int = self.prepare_autoformer_inputs_dict(__lowercase ) return config, inputs_dict def UpperCamelCase__ ( self ) -> Optional[int]: a_ : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def UpperCamelCase__ ( self , _lowercase , _lowercase ) -> int: a_ : int = AutoformerModel(config=__lowercase ).to(__lowercase ).eval() a_ : int = model(**__lowercase ) a_ : Optional[Any] = outputs.encoder_last_hidden_state a_ : Tuple = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: a_ : int = model.get_encoder() encoder.save_pretrained(__lowercase ) a_ : List[str] = AutoformerEncoder.from_pretrained(__lowercase ).to(__lowercase ) a_ : List[Any] = model.create_network_inputs(**__lowercase ) a_ : Tuple = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) a_ : Tuple = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) a_ : str = encoder(inputs_embeds=__lowercase )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) a_ : int = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) a_ : int = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) a_ : int = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) a_ : Dict = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: a_ : Optional[Any] = model.get_decoder() decoder.save_pretrained(__lowercase ) a_ : List[Any] = AutoformerDecoder.from_pretrained(__lowercase ).to(__lowercase ) a_ : Optional[int] = decoder( trend=__lowercase , inputs_embeds=__lowercase , encoder_hidden_states=__lowercase , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class A__(lowercase__, lowercase__, unittest.TestCase ): """simple docstring""" _A : Dict = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () _A : Any = (AutoformerForPrediction,) if is_torch_available() else () _A : Dict = {'feature-extraction': AutoformerModel} if is_torch_available() else {} _A : Optional[int] = False _A : List[str] = False _A : str = False _A : Optional[int] = False _A : Union[str, Any] = False _A : Tuple = False def UpperCamelCase__ ( self ) -> List[str]: a_ : Tuple = AutoformerModelTester(self ) a_ : Dict = ConfigTester(self , config_class=__lowercase , has_text_modality=__lowercase ) def UpperCamelCase__ ( self ) -> Tuple: self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) -> str: a_ : List[str] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: a_ : Any = model_class(__lowercase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowercase ) a_ : Union[str, Any] = model_class.from_pretrained(__lowercase , output_loading_info=__lowercase ) self.assertEqual(info["""missing_keys"""] , [] ) def UpperCamelCase__ ( self ) -> int: a_ : str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__lowercase ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def UpperCamelCase__ ( self ) -> List[Any]: pass def UpperCamelCase__ ( self ) -> Optional[Any]: a_ : Any = inspect.signature(getattr(__lowercase , """forward""" ) ) # The main input is the name of the argument after `self` a_ : Optional[int] = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , __lowercase ) def UpperCamelCase__ ( self ) -> Union[str, Any]: a_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : Tuple = model_class(__lowercase ) a_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a_ : Tuple = [*signature.parameters.keys()] a_ : Tuple = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(__lowercase )] , __lowercase ) def UpperCamelCase__ ( self ) -> Optional[int]: a_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() a_ : Any = True a_ : List[str] = getattr(self.model_tester , """seq_length""" , __lowercase ) a_ : Optional[int] = getattr(self.model_tester , """decoder_seq_length""" , __lowercase ) a_ : Dict = getattr(self.model_tester , """encoder_seq_length""" , __lowercase ) a_ : Union[str, Any] = getattr(self.model_tester , """d_model""" , __lowercase ) a_ : List[Any] = getattr(self.model_tester , """num_attention_heads""" , __lowercase ) a_ : List[Any] = d_model // num_attention_heads for model_class in self.all_model_classes: a_ : Dict = True a_ : str = False a_ : Optional[Any] = True a_ : Union[str, Any] = model_class(__lowercase ) model.to(__lowercase ) model.eval() with torch.no_grad(): a_ : int = model(**self._prepare_for_class(__lowercase , __lowercase ) ) a_ : List[Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] a_ : int = True a_ : int = model_class(__lowercase ) model.to(__lowercase ) model.eval() with torch.no_grad(): a_ : List[str] = model(**self._prepare_for_class(__lowercase , __lowercase ) ) a_ : Optional[int] = outputs.encoder_attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) a_ : List[str] = len(__lowercase ) a_ : Union[str, Any] = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(__lowercase , __lowercase ) # decoder attentions a_ : List[Any] = outputs.decoder_attentions self.assertIsInstance(__lowercase , (list, tuple) ) self.assertEqual(len(__lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions a_ : Tuple = outputs.cross_attentions self.assertIsInstance(__lowercase , (list, tuple) ) self.assertEqual(len(__lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine a_ : List[Any] = True a_ : int = True a_ : List[str] = model_class(__lowercase ) model.to(__lowercase ) model.eval() with torch.no_grad(): a_ : int = model(**self._prepare_for_class(__lowercase , __lowercase ) ) self.assertEqual(out_len + 2 , len(__lowercase ) ) a_ : Dict = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def UpperCamelCase__ ( self ) -> Optional[int]: super().test_retain_grad_hidden_states_attentions() def _UpperCAmelCase ( a__="train-batch.pt"): '''simple docstring''' a_ : str = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=__lowerCamelCase , repo_type="""dataset""") a_ : Optional[Any] = torch.load(__lowerCamelCase , map_location=__lowerCamelCase) return batch @require_torch @slow class A__(unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ) -> List[Any]: a_ : Dict = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(__lowercase ) a_ : Dict = prepare_batch() with torch.no_grad(): a_ : int = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , future_values=batch["""future_values"""] , future_time_features=batch["""future_time_features"""] , )[0] a_ : Dict = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , __lowercase ) a_ : Optional[int] = torch.tensor( [[0.3_5_9_3, -1.3_3_9_8, 0.6_3_3_0], [0.2_2_7_9, 1.5_3_9_6, -0.1_7_9_2], [0.0_4_5_0, 1.3_2_2_5, -0.2_3_3_5]] , device=__lowercase ) self.assertTrue(torch.allclose(output[0, :3, :3] , __lowercase , atol=__lowercase ) ) def UpperCamelCase__ ( self ) -> Union[str, Any]: a_ : Union[str, Any] = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(__lowercase ) a_ : List[str] = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): a_ : Any = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , ).encoder_last_hidden_state a_ : Optional[int] = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , __lowercase ) a_ : Dict = torch.tensor( [[-0.0_7_3_4, -0.9_0_3_6, 0.8_3_5_8], [4.7_1_8_6, 2.4_1_1_3, 1.9_5_8_1], [1.7_9_5_3, 2.3_5_5_8, 1.2_9_7_0]] , device=__lowercase ) self.assertTrue(torch.allclose(output[0, :3, :3] , __lowercase , atol=__lowercase ) ) def UpperCamelCase__ ( self ) -> Optional[int]: a_ : Union[str, Any] = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(__lowercase ) a_ : Optional[int] = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): a_ : int = model.generate( static_categorical_features=batch["""static_categorical_features"""] , past_time_features=batch["""past_time_features"""] , past_values=batch["""past_values"""] , future_time_features=batch["""future_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , ) a_ : Optional[Any] = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , __lowercase ) a_ : Optional[Any] = torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=__lowercase ) a_ : List[str] = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , __lowercase , rtol=1e-1 ) )
540
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Any=13 , __lowercase : Optional[int]=7 , __lowercase : str=True , __lowercase : Optional[Any]=True , __lowercase : int=True , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : int=5 , __lowercase : Tuple=4 , __lowercase : str=37 , __lowercase : Optional[int]="gelu" , __lowercase : Tuple=0.1 , __lowercase : str=0.1 , __lowercase : Dict=512 , __lowercase : List[Any]=16 , __lowercase : Dict=2 , __lowercase : Union[str, Any]=0.02 , __lowercase : Dict=4 , ) -> int: __UpperCAmelCase : Dict = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : str = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Dict = use_token_type_ids __UpperCAmelCase : Optional[int] = use_labels __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : str = type_sequence_label_size __UpperCAmelCase : List[str] = initializer_range __UpperCAmelCase : Tuple = num_choices def UpperCAmelCase ( self : Dict ) -> Tuple: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : List[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase ( self : Tuple ) -> List[Any]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : List[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase ( self : Any ) -> List[str]: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[int] = config_and_inputs __UpperCAmelCase : int = True __UpperCAmelCase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[Any] = True a : List[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase : List[str] = FlaxRobertaModelTester(self ) @slow def UpperCAmelCase ( self : str ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase : Union[str, Any] = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowercase ) __UpperCAmelCase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowercase )
63
0
'''simple docstring''' import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def __snake_case ( lowerCamelCase_ : int , lowerCamelCase_ : Any , lowerCamelCase_ : Union[str, Any] ): '''simple docstring''' __magic_name__ = AlbertConfig.from_json_file(__lowerCamelCase ) print(F'Building PyTorch model from configuration: {config}' ) __magic_name__ = AlbertForPreTraining(__lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_albert(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , __lowerCamelCase ) if __name__ == "__main__": __magic_name__ : Optional[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--albert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained ALBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __magic_name__ : Union[str, Any] =parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
664
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging a : Optional[int] = logging.get_logger(__name__) class a ( lowercase__ ): """simple docstring""" a : Tuple = 'linear' a : int = 'cosine' a : Optional[Any] = 'cosine_with_restarts' a : Dict = 'polynomial' a : Tuple = 'constant' a : Dict = 'constant_with_warmup' a : Any = 'piecewise_constant' def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int = -1 ): return LambdaLR(__lowerCamelCase , lambda __lowerCamelCase : 1 , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : str , __lowerCamelCase : int = -1 ): __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Tuple = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __UpperCAmelCase , __UpperCAmelCase : List[str] = rule_str.split(""":""" ) __UpperCAmelCase : Any = int(__lowerCamelCase ) __UpperCAmelCase : List[str] = float(__lowerCamelCase ) __UpperCAmelCase : int = value __UpperCAmelCase : Any = float(rule_list[-1] ) def create_rules_function(__lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): def rule_func(__lowerCamelCase : int ) -> float: __UpperCAmelCase : Tuple = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=-1 ): def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float = 0.5 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Dict ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Tuple = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optimizer , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int = 1 , __lowerCamelCase : int = -1 ): def lr_lambda(__lowerCamelCase : Union[str, Any] ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) __UpperCAmelCase : Union[str, Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any]=1E-7 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : int=-1 ): __UpperCAmelCase : Tuple = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(f"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowerCamelCase : int ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __UpperCAmelCase : Optional[Any] = lr_init - lr_end __UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps __UpperCAmelCase : int = 1 - (current_step - num_warmup_steps) / decay_steps __UpperCAmelCase : Optional[int] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a : int = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCamelCase__ ( __lowerCamelCase : Union[str, SchedulerType] , __lowerCamelCase : Optimizer , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 1 , __lowerCamelCase : float = 1.0 , __lowerCamelCase : int = -1 , ): __UpperCAmelCase : Union[str, Any] = SchedulerType(__lowerCamelCase ) __UpperCAmelCase : int = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(f"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(f"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
63
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowercase ( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = StableDiffusionXLImgaImgPipeline __SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} __SCREAMING_SNAKE_CASE = PipelineTesterMixin.required_optional_params - {'latents'} __SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __SCREAMING_SNAKE_CASE = IMAGE_TO_IMAGE_IMAGE_PARAMS __SCREAMING_SNAKE_CASE = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , attention_head_dim=(2, 4) , use_linear_projection=__lowercase , addition_embed_type='text_time' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) UpperCAmelCase__ = EulerDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , steps_offset=1 , beta_schedule='scaled_linear' , timestep_spacing='leading' , ) torch.manual_seed(0 ) UpperCAmelCase__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=32 , ) UpperCAmelCase__ = CLIPTextModel(__lowercase ) UpperCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=__lowercase ) UpperCAmelCase__ = CLIPTextModelWithProjection(__lowercase ) UpperCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=__lowercase ) UpperCAmelCase__ = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """text_encoder_2""": text_encoder_a, """tokenizer_2""": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def UpperCamelCase__ (self , __a , __a=0 ) -> str: """simple docstring""" UpperCAmelCase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase ) ).to(__lowercase ) UpperCAmelCase__ = image / 2 + 0.5 if str(__lowercase ).startswith('mps' ): UpperCAmelCase__ = torch.manual_seed(__lowercase ) else: UpperCAmelCase__ = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) UpperCAmelCase__ = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 5.0, """output_type""": """numpy""", """strength""": 0.75, } return inputs def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = StableDiffusionXLImgaImgPipeline(**__lowercase ) UpperCAmelCase__ = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) UpperCAmelCase__ = self.get_dummy_inputs(__lowercase ) UpperCAmelCase__ = sd_pipe(**__lowercase ).images UpperCAmelCase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCAmelCase__ = np.array([0.46_56, 0.48_40, 0.44_39, 0.66_98, 0.55_74, 0.45_24, 0.57_99, 0.59_43, 0.51_65] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = StableDiffusionXLImgaImgPipeline(**__lowercase ) UpperCAmelCase__ = sd_pipe.to(__lowercase ) UpperCAmelCase__ = sd_pipe.to(__lowercase ) sd_pipe.set_progress_bar_config(disable=__lowercase ) # forward without prompt embeds UpperCAmelCase__ = self.get_dummy_inputs(__lowercase ) UpperCAmelCase__ = 3 * ["""this is a negative prompt"""] UpperCAmelCase__ = negative_prompt UpperCAmelCase__ = 3 * [inputs["""prompt"""]] UpperCAmelCase__ = sd_pipe(**__lowercase ) UpperCAmelCase__ = output.images[0, -3:, -3:, -1] # forward with prompt embeds UpperCAmelCase__ = self.get_dummy_inputs(__lowercase ) UpperCAmelCase__ = 3 * ["""this is a negative prompt"""] UpperCAmelCase__ = 3 * [inputs.pop('prompt' )] ( UpperCAmelCase__ ) = sd_pipe.encode_prompt(__lowercase , negative_prompt=__lowercase ) UpperCAmelCase__ = sd_pipe( **__lowercase , prompt_embeds=__lowercase , negative_prompt_embeds=__lowercase , pooled_prompt_embeds=__lowercase , negative_pooled_prompt_embeds=__lowercase , ) UpperCAmelCase__ = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class lowercase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ (self , __a , __a="cpu" , __a=torch.floataa , __a=0 ) -> Any: """simple docstring""" UpperCAmelCase__ = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) UpperCAmelCase__ = np.random.RandomState(__lowercase ).standard_normal((1, 4, 64, 64) ) UpperCAmelCase__ = torch.from_numpy(__lowercase ).to(device=__lowercase , dtype=__lowercase ) UpperCAmelCase__ = { """prompt""": """a photograph of an astronaut riding a horse""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = DiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base' ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) UpperCAmelCase__ = self.get_inputs(__lowercase ) UpperCAmelCase__ = pipe(**__lowercase ).images UpperCAmelCase__ = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) UpperCAmelCase__ = np.array([0.4_94_93, 0.4_78_96, 0.4_07_98, 0.5_42_14, 0.5_32_12, 0.4_82_02, 0.4_76_56, 0.4_63_29, 0.4_85_06] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
146
from math import pi, sqrt def lowerCamelCase__ ( __lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCamelCase__ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() a : Optional[int] = 1.0 while num: a : List[str] = float(input("Gamma of: ")) print(f"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
63
0
'''simple docstring''' import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed lowerCAmelCase__ = logging.getLogger(__name__) def _A ( A__=2 , A__=3 , A__=16 , A__ = 10 , A__ = 2 ): """simple docstring""" def get_dataset(A__ ): __lowercase = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(__lowerCamelCase , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) __lowercase = get_dataset(__lowerCamelCase ) __lowercase = get_dataset(__lowerCamelCase ) __lowercase = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) __lowercase = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) return (train_dataloader, valid_dataloader) def _A ( A__ , A__ , A__ , A__ , A__ , A__=None ): """simple docstring""" __lowercase = [] for epoch in range(__lowerCamelCase ): # Train quickly model.train() for batch in dataloader: __lowercase = batch __lowercase = model(__lowerCamelCase ) __lowercase = torch.nn.functional.mse_loss(__lowerCamelCase , __lowerCamelCase ) accelerator.backward(__lowerCamelCase ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class lowercase_ (nn.Module ): """simple docstring""" def __init__( self : Union[str, Any] ): super().__init__() __lowercase = nn.Parameter(torch.randn(1 ) ) __lowercase = nn.Parameter(torch.randn(1 ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[str] ): return x * self.a + self.b class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[str] ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(total_limit=1 ,project_dir=__lowercase ,automatic_checkpoint_naming=__lowercase ) # Train baseline __lowercase = Accelerator(project_config=__lowercase ) __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) ,1 ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = dummy_dataloaders() # Train baseline __lowercase = Accelerator() __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save initial __lowercase = os.path.join(__lowercase ,'''initial''' ) accelerator.save_state(__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() __lowercase = train(3 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() # Train partially set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = dummy_dataloaders() __lowercase = Accelerator() __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ) accelerator.load_state(__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) __lowercase = train(2 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save everything __lowercase = os.path.join(__lowercase ,'''checkpoint''' ) accelerator.save_state(__lowercase ) # Load everything back in and make sure all states work accelerator.load_state(__lowercase ) test_rands += train(1 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : int ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=__lowercase ) # Train baseline __lowercase = Accelerator(project_dir=__lowercase ,project_config=__lowercase ) __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save initial accelerator.save_state() (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() __lowercase = train(3 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() # Train partially set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(iteration=1 ,automatic_checkpoint_naming=__lowercase ) __lowercase = Accelerator(project_dir=__lowercase ,project_config=__lowercase ) __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ) accelerator.load_state(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_0''' ) ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) __lowercase = train(2 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_1''' ) ) test_rands += train(1 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) (__lowercase) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) self.assertEqual(__lowercase ,__lowercase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = torch.tensor([1, 2, 3] ) __lowercase = torch.tensor([2, 3, 4] ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(net.parameters() ) __lowercase = Accelerator() with self.assertRaises(__lowercase ) as ve: accelerator.register_for_checkpointing(__lowercase ,__lowercase ,__lowercase ,__lowercase ) __lowercase = str(ve.exception ) self.assertTrue('''Item at index 0''' in message ) self.assertTrue('''Item at index 1''' in message ) self.assertFalse('''Item at index 2''' in message ) self.assertFalse('''Item at index 3''' in message ) def SCREAMING_SNAKE_CASE ( self : Dict ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) __lowercase = torch.optim.lr_scheduler.StepLR(__lowercase ,step_size=1 ,gamma=0.9_9 ) __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=__lowercase ) # Train baseline __lowercase = Accelerator(project_dir=__lowercase ,project_config=__lowercase ) __lowercase = accelerator.prepare( __lowercase ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) # Save initial accelerator.save_state() __lowercase = scheduler.state_dict() train(3 ,__lowercase ,__lowercase ,__lowercase ,__lowercase ,__lowercase ) self.assertNotEqual(__lowercase ,scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_0''' ) ) self.assertEqual(__lowercase ,scheduler.state_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) __lowercase = DummyModel() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=__lowercase ,total_limit=2 ) # Train baseline __lowercase = Accelerator(project_dir=__lowercase ,project_config=__lowercase ) __lowercase = accelerator.prepare(__lowercase ) # Save 3 states: for _ in range(1_1 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_9''' ) ) ) self.assertTrue(os.path.exists(os.path.join(__lowercase ,'''checkpoints''' ,'''checkpoint_10''' ) ) ) @require_cuda def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ["""torchrun""", F"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] execute_subprocess_async(__lowercase ,env=os.environ.copy() ) if __name__ == "__main__": lowerCAmelCase__ = "/tmp/accelerate/state_checkpointing" lowerCAmelCase__ = DummyModel() lowerCAmelCase__ = torch.optim.Adam(params=model.parameters(), lr=1e-3) lowerCAmelCase__ = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) lowerCAmelCase__ = dummy_dataloaders() lowerCAmelCase__ = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline lowerCAmelCase__ = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='''no''') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) lowerCAmelCase__ = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) lowerCAmelCase__ = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: lowerCAmelCase__ = group["params"][0].device break assert param_device.type == accelerator.device.type lowerCAmelCase__ = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''cpu''') for group in optimizer.param_groups: lowerCAmelCase__ = group["params"][0].device break assert ( param_device.type == torch.device('''cpu''').type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''on_device''') for group in optimizer.param_groups: lowerCAmelCase__ = group["params"][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='''Unsupported optimizer map location passed'''): accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''invalid''') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
41
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class a : """simple docstring""" a : int a : Node | None = None a : Node | None = None def lowerCamelCase__ ( ): __UpperCAmelCase : Tuple = Node(1 ) __UpperCAmelCase : int = Node(2 ) __UpperCAmelCase : Optional[Any] = Node(3 ) __UpperCAmelCase : Dict = Node(4 ) __UpperCAmelCase : Tuple = Node(5 ) return tree def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCamelCase__ ( __lowerCamelCase : Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCamelCase__ ( __lowerCamelCase : Node | None ): __UpperCAmelCase : list[Any] = [] if root is None: return output __UpperCAmelCase : Tuple = deque([root] ) while process_queue: __UpperCAmelCase : Optional[Any] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None , __lowerCamelCase : int ): __UpperCAmelCase : list[Any] = [] def populate_output(__lowerCamelCase : Node | None , __lowerCamelCase : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__lowerCamelCase , __lowerCamelCase ) return output def lowerCamelCase__ ( __lowerCamelCase : Node | None ): if root is None: return [] __UpperCAmelCase : list[Sequence[Node | None]] = [] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : int = height(__lowerCamelCase ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : Optional[int] = 0 return output def lowerCamelCase__ ( ): # Main function for testing. __UpperCAmelCase : List[Any] = make_tree() print(f"""In-order Traversal: {inorder(__lowerCamelCase )}""" ) print(f"""Pre-order Traversal: {preorder(__lowerCamelCase )}""" ) print(f"""Post-order Traversal: {postorder(__lowerCamelCase )}""" , """\n""" ) print(f"""Height of Tree: {height(__lowerCamelCase )}""" , """\n""" ) print("""Complete Level Order Traversal: """ ) print(level_order(__lowerCamelCase ) , """\n""" ) print("""Level-wise order Traversal: """ ) for level in range(1 , height(__lowerCamelCase ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(__lowerCamelCase , level=__lowerCamelCase ) ) print("""\nZigZag order Traversal: """ ) print(zigzag(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
63
0
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE :Tuple = { "configuration_xmod": [ "XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP", "XmodConfig", "XmodOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE :Optional[Any] = [ "XMOD_PRETRAINED_MODEL_ARCHIVE_LIST", "XmodForCausalLM", "XmodForMaskedLM", "XmodForMultipleChoice", "XmodForQuestionAnswering", "XmodForSequenceClassification", "XmodForTokenClassification", "XmodModel", "XmodPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE :Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
283
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Optional[int] = GPTSanJapaneseTokenizer a : Optional[Any] = False a : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off __UpperCAmelCase : Tuple = ["""こん""", """こんに""", """にちは""", """ばんは""", """世界,㔺界""", """、""", """。""", """<BR>""", """<SP>""", """<TAB>""", """<URL>""", """<EMAIL>""", """<TEL>""", """<DATE>""", """<PRICE>""", """<BLOCK>""", """<KIGOU>""", """<U2000U2BFF>""", """<|emoji1|>""", """<unk>""", """<|bagoftoken|>""", """<|endoftext|>"""] # fmt: on __UpperCAmelCase : Dict = {"""emoji""": {"""\ud83d\ude00""": """<|emoji1|>"""}, """emoji_inv""": {"""<|emoji1|>""": """\ud83d\ude00"""}} # 😀 __UpperCAmelCase : Dict = {"""unk_token""": """<unk>"""} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""emoji_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.emoji_file , """w""" ) as emoji_writer: emoji_writer.write(json.dumps(__lowercase ) ) def UpperCAmelCase ( self : Tuple , **__lowercase : int ) -> Any: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__lowercase ) def UpperCAmelCase ( self : str , __lowercase : Union[str, Any] ) -> Any: __UpperCAmelCase : Any = """こんにちは、世界。 \nこんばんは、㔺界。😀""" __UpperCAmelCase : int = """こんにちは、世界。 \nこんばんは、世界。😀""" return input_text, output_text def UpperCAmelCase ( self : List[Any] , __lowercase : Optional[int] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase : int = self.get_input_output_texts(__lowercase ) __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __UpperCAmelCase : Dict = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) return text, ids def UpperCAmelCase ( self : int ) -> Optional[Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: pass # TODO add if relevant def UpperCAmelCase ( self : Dict ) -> Tuple: pass # TODO add if relevant def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : List[str] = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。 こんばんは、㔺界。""" __UpperCAmelCase : Dict = ["""こん""", """にちは""", """、""", """世界""", """。""", """<SP>""", """こん""", """ばんは""", """、""", """㔺界""", """。"""] __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids without special tokens __UpperCAmelCase : List[str] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # Testing conversion to ids with special tokens __UpperCAmelCase : List[Any] = tokens + [tokenizer.unk_token] __UpperCAmelCase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : int = self.get_tokenizer() # Testing tokenization __UpperCAmelCase : Tuple = """こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。""" __UpperCAmelCase : int = """こんにちは、、、、世界。こんばんは、、、、世界。""" __UpperCAmelCase : Tuple = tokenizer.encode(__lowercase ) __UpperCAmelCase : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : int ) -> Optional[int]: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : List[Any] = """こんにちは、世界。""" __UpperCAmelCase : Optional[int] = """こんばんは、㔺界。😀""" __UpperCAmelCase : List[Any] = """こんにちは、世界。こんばんは、世界。😀""" __UpperCAmelCase : List[str] = tokenizer.encode(prefix_text + input_text ) __UpperCAmelCase : List[Any] = tokenizer.encode("""""" , prefix_text=prefix_text + input_text ) __UpperCAmelCase : Any = tokenizer.encode(__lowercase , prefix_text=__lowercase ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowercase ) __UpperCAmelCase : Any = tokenizer.decode(__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : int = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) # Testing tokenization __UpperCAmelCase : int = """こんにちは、世界。""" __UpperCAmelCase : List[Any] = """こんばんは、㔺界。😀""" __UpperCAmelCase : Union[str, Any] = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : int = len(tokenizer.encode(__lowercase ) ) - 2 __UpperCAmelCase : List[Any] = [1] + [0] * (len_prefix + len_text + 1) __UpperCAmelCase : Union[str, Any] = [1] * (len_prefix + len_text + 1) + [0] __UpperCAmelCase : List[Any] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCAmelCase : Union[str, Any] = tokenizer(prefix_text + input_text ).token_type_ids __UpperCAmelCase : Optional[Any] = tokenizer("""""" , prefix_text=prefix_text + input_text ).token_type_ids __UpperCAmelCase : Tuple = tokenizer(__lowercase , prefix_text=__lowercase ).token_type_ids self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def UpperCAmelCase ( self : List[str] ) -> int: __UpperCAmelCase : Tuple = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""あンいワ""" ) __UpperCAmelCase : Tuple = tokenizer.encode("""""" , prefix_text="""あンいワ""" ) __UpperCAmelCase : Optional[int] = tokenizer.encode("""いワ""" , prefix_text="""あン""" ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertEqual(tokenizer.decode(__lowercase ) , tokenizer.decode(__lowercase ) ) self.assertNotEqual(__lowercase , __lowercase ) self.assertNotEqual(__lowercase , __lowercase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def UpperCAmelCase ( self : List[Any] ) -> List[str]: __UpperCAmelCase : Any = self.tokenizer_class.from_pretrained("""Tanrei/GPTSAN-japanese""" ) __UpperCAmelCase : List[Any] = [["""武田信玄""", """は、"""], ["""織田信長""", """の配下の、"""]] __UpperCAmelCase : int = tokenizer(__lowercase , padding=__lowercase ) __UpperCAmelCase : Optional[Any] = tokenizer.batch_encode_plus(__lowercase , padding=__lowercase ) # fmt: off __UpperCAmelCase : Optional[int] = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCAmelCase : Tuple = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCAmelCase : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __lowercase ) self.assertListEqual(x_token.token_type_ids , __lowercase ) self.assertListEqual(x_token.attention_mask , __lowercase ) self.assertListEqual(x_token_a.input_ids , __lowercase ) self.assertListEqual(x_token_a.token_type_ids , __lowercase ) self.assertListEqual(x_token_a.attention_mask , __lowercase ) def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def UpperCAmelCase ( self : Any ) -> int: # tokenizer has no padding token pass
63
0
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __a : def __init__( self : Union[str, Any] , UpperCAmelCase : str = "cpu" , UpperCAmelCase : str = "openai/clip-vit-large-patch14" ): lowerCAmelCase_ : List[str] = device lowerCAmelCase_ : Any = CLIPTokenizerFast.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[int] = [0.4814_5466, 0.457_8275, 0.4082_1073] lowerCAmelCase_ : List[str] = [0.2686_2954, 0.2613_0258, 0.2757_7711] lowerCAmelCase_ : Dict = torchvision.transforms.Normalize(self.image_mean , self.image_std ) lowerCAmelCase_ : Tuple = torchvision.transforms.Resize(2_24 ) lowerCAmelCase_ : Union[str, Any] = torchvision.transforms.CenterCrop(2_24 ) def A ( self : int , UpperCAmelCase : Dict ): lowerCAmelCase_ : Tuple = self.resize(__lowercase ) lowerCAmelCase_ : Union[str, Any] = self.center_crop(__lowercase ) lowerCAmelCase_ : Any = self.normalize(__lowercase ) return images def __call__( self : List[Any] , UpperCAmelCase : Any=None , UpperCAmelCase : str=None , **UpperCAmelCase : int ): lowerCAmelCase_ : Union[str, Any] = self.tokenizer(text=__lowercase , **__lowercase ) lowerCAmelCase_ : str = self.preprocess_img(__lowercase ) lowerCAmelCase_ : Any = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class __a ( nn.Module ): def __init__( self : List[str] , UpperCAmelCase : Optional[int]=10 , UpperCAmelCase : Optional[int]=0.01 , UpperCAmelCase : str=None , UpperCAmelCase : str=None , UpperCAmelCase : List[Any]=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : List[str]=False , UpperCAmelCase : Dict=True , UpperCAmelCase : Optional[Any]="image" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Union[str, Any]=False , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=False , ): super().__init__() lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : Optional[Any] = device if device else get_device() if vqgan: lowerCAmelCase_ : Dict = vqgan else: lowerCAmelCase_ : Optional[int] = load_vqgan(self.device , conf_path=__lowercase , ckpt_path=__lowercase ) self.vqgan.eval() if clip: lowerCAmelCase_ : Optional[int] = clip else: lowerCAmelCase_ : Dict = CLIPModel.from_pretrained("""openai/clip-vit-base-patch32""" ) self.clip.to(self.device ) lowerCAmelCase_ : Optional[Any] = ProcessorGradientFlow(device=self.device ) lowerCAmelCase_ : int = iterations lowerCAmelCase_ : Union[str, Any] = lr lowerCAmelCase_ : Dict = log lowerCAmelCase_ : Tuple = make_grid lowerCAmelCase_ : List[str] = return_val lowerCAmelCase_ : Any = quantize lowerCAmelCase_ : str = self.vqgan.decoder.z_shape def A ( self : List[str] , UpperCAmelCase : Any=None , UpperCAmelCase : Any=None , UpperCAmelCase : Union[str, Any]=5 , UpperCAmelCase : str=True ): lowerCAmelCase_ : Dict = [] if output_path is None: lowerCAmelCase_ : Dict = """./animation.gif""" if input_path is None: lowerCAmelCase_ : str = self.save_path lowerCAmelCase_ : Any = sorted(glob(input_path + """/*""" ) ) if not len(__lowercase ): raise ValueError( """No images found in save path, aborting (did you pass save_intermediate=True to the generate""" """ function?)""" ) if len(__lowercase ) == 1: print("""Only one image found in save path, (did you pass save_intermediate=True to the generate function?)""" ) lowerCAmelCase_ : Tuple = total_duration / len(__lowercase ) lowerCAmelCase_ : Union[str, Any] = [frame_duration] * len(__lowercase ) if extend_frames: lowerCAmelCase_ : Any = 1.5 lowerCAmelCase_ : Tuple = 3 for file_name in paths: if file_name.endswith(""".png""" ): images.append(imageio.imread(__lowercase ) ) imageio.mimsave(__lowercase , __lowercase , duration=__lowercase ) print(F'gif saved to {output_path}' ) def A ( self : Optional[Any] , UpperCAmelCase : Dict=None , UpperCAmelCase : List[Any]=None ): if not (path or img): raise ValueError("""Input either path or tensor""" ) if img is not None: raise NotImplementedError lowerCAmelCase_ : Optional[Any] = preprocess(Image.open(__lowercase ) , target_image_size=2_56 ).to(self.device ) lowerCAmelCase_ : str = preprocess_vqgan(__lowercase ) lowerCAmelCase_ : List[Any] = self.vqgan.encode(__lowercase ) return z def A ( self : str , UpperCAmelCase : List[str] ): lowerCAmelCase_ : Any = self.latent.detach().requires_grad_() lowerCAmelCase_ : Any = base_latent + transform_vector if self.quantize: lowerCAmelCase_ : Union[str, Any] = self.vqgan.quantize(__lowercase ) else: lowerCAmelCase_ : Tuple = trans_latent return self.vqgan.decode(__lowercase ) def A ( self : List[str] , UpperCAmelCase : str , UpperCAmelCase : Dict , UpperCAmelCase : List[str]=None ): lowerCAmelCase_ : int = self.clip_preprocessor(text=__lowercase , images=__lowercase , return_tensors="""pt""" , padding=__lowercase ) lowerCAmelCase_ : int = self.clip(**__lowercase ) lowerCAmelCase_ : List[Any] = clip_outputs.logits_per_image if weights is not None: lowerCAmelCase_ : Union[str, Any] = similarity_logits * weights return similarity_logits.sum() def A ( self : Any , UpperCAmelCase : Any , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int] ): lowerCAmelCase_ : Union[str, Any] = self._get_clip_similarity(pos_prompts["""prompts"""] , __lowercase , weights=(1 / pos_prompts["""weights"""]) ) if neg_prompts: lowerCAmelCase_ : Union[str, Any] = self._get_clip_similarity(neg_prompts["""prompts"""] , __lowercase , weights=neg_prompts["""weights"""] ) else: lowerCAmelCase_ : str = torch.tensor([1] , device=self.device ) lowerCAmelCase_ : List[str] = -torch.log(__lowercase ) + torch.log(__lowercase ) return loss def A ( self : Optional[Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple ): lowerCAmelCase_ : Optional[int] = torch.randn_like(self.latent , requires_grad=__lowercase , device=self.device ) lowerCAmelCase_ : Tuple = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() lowerCAmelCase_ : Optional[Any] = self._add_vector(__lowercase ) lowerCAmelCase_ : List[Any] = loop_post_process(__lowercase ) lowerCAmelCase_ : List[Any] = self._get_CLIP_loss(__lowercase , __lowercase , __lowercase ) print("""CLIP loss""" , __lowercase ) if self.log: wandb.log({"""CLIP Loss""": clip_loss} ) clip_loss.backward(retain_graph=__lowercase ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def A ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int] ): wandb.init(reinit=__lowercase , project="""face-editor""" ) wandb.config.update({"""Positive Prompts""": positive_prompts} ) wandb.config.update({"""Negative Prompts""": negative_prompts} ) wandb.config.update({"""lr""": self.lr, """iterations""": self.iterations} ) if image_path: lowerCAmelCase_ : Optional[int] = Image.open(__lowercase ) lowerCAmelCase_ : Union[str, Any] = image.resize((2_56, 2_56) ) wandb.log("""Original Image""" , wandb.Image(__lowercase ) ) def A ( self : str , UpperCAmelCase : Tuple ): if not prompts: return [] lowerCAmelCase_ : str = [] lowerCAmelCase_ : str = [] if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Any = [prompt.strip() for prompt in prompts.split("""|""" )] for prompt in prompts: if isinstance(__lowercase , (tuple, list) ): lowerCAmelCase_ : str = prompt[0] lowerCAmelCase_ : Optional[int] = float(prompt[1] ) elif ":" in prompt: lowerCAmelCase_ : Tuple = prompt.split(""":""" ) lowerCAmelCase_ : Union[str, Any] = float(__lowercase ) else: lowerCAmelCase_ : str = prompt lowerCAmelCase_ : List[Any] = 1.0 processed_prompts.append(__lowercase ) weights.append(__lowercase ) return { "prompts": processed_prompts, "weights": torch.tensor(__lowercase , device=self.device ), } def A ( self : Optional[int] , UpperCAmelCase : Tuple , UpperCAmelCase : Dict=None , UpperCAmelCase : Optional[int]=None , UpperCAmelCase : List[str]=True , UpperCAmelCase : Dict=False , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]=None , ): if image_path: lowerCAmelCase_ : Optional[Any] = self._get_latent(__lowercase ) else: lowerCAmelCase_ : int = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(__lowercase , __lowercase , __lowercase ) assert pos_prompts, "You must provide at least one positive prompt." lowerCAmelCase_ : Dict = self.process_prompts(__lowercase ) lowerCAmelCase_ : Optional[Any] = self.process_prompts(__lowercase ) if save_final and save_path is None: lowerCAmelCase_ : Optional[Any] = os.path.join("""./outputs/""" , """_""".join(pos_prompts["""prompts"""] ) ) if not os.path.exists(__lowercase ): os.makedirs(__lowercase ) else: lowerCAmelCase_ : Dict = save_path + """_""" + get_timestamp() os.makedirs(__lowercase ) lowerCAmelCase_ : List[str] = save_path lowerCAmelCase_ : Optional[int] = self.vqgan.decode(self.latent )[0] if show_intermediate: print("""Original Image""" ) show_pil(custom_to_pil(__lowercase ) ) lowerCAmelCase_ : Dict = loop_post_process(__lowercase ) for iter, transformed_img in enumerate(self._optimize_CLIP(__lowercase , __lowercase , __lowercase ) ): if show_intermediate: show_pil(__lowercase ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}.png' ) ) if self.log: wandb.log({"""Image""": wandb.Image(__lowercase )} ) if show_final: show_pil(__lowercase ) if save_final: transformed_img.save(os.path.join(self.save_path , F'iter_{iter:03d}_final.png' ) )
600
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a : Dict = logging.get_logger(__name__) @dataclass class a ( lowercase__ ): """simple docstring""" a : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : List[Any] , **__lowercase : Dict ) -> Tuple: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase : List[Any] = deprecated_arg[3:] setattr(self , __lowercase , not kwargs.pop(__lowercase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __UpperCAmelCase : str = kwargs.pop("""torchscript""" , self.torchscript ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) __UpperCAmelCase : Optional[Any] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**__lowercase ) a : bool = field(default=lowercase__ , metadata={'help': 'Trace the models using torchscript'} ) a : bool = field(default=lowercase__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) a : str = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def UpperCAmelCase ( self : Any ) -> Tuple["torch.device", int]: requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: __UpperCAmelCase : str = torch.device("""cpu""" ) __UpperCAmelCase : int = 0 elif is_torch_tpu_available(): __UpperCAmelCase : Tuple = xm.xla_device() __UpperCAmelCase : int = 0 else: __UpperCAmelCase : Dict = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) __UpperCAmelCase : Optional[int] = torch.cuda.device_count() return device, n_gpu @property def UpperCAmelCase ( self : Optional[Any] ) -> str: return is_torch_tpu_available() and self.tpu @property def UpperCAmelCase ( self : List[str] ) -> int: requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCAmelCase ( self : int ) -> "torch.device": requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def UpperCAmelCase ( self : int ) -> List[Any]: requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def UpperCAmelCase ( self : Tuple ) -> List[str]: return self.n_gpu > 0
63
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowerCamelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
6
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict ): assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __UpperCAmelCase : str = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : Any ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __UpperCAmelCase : Any = features.copy() if features else default_expected_features __UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=__lowerCamelCase , cache_dir=__lowerCamelCase ).read() _check_sql_dataset(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] ): with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: __UpperCAmelCase : Dict = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): __UpperCAmelCase : Optional[int] = tmp_path / """cache""" __UpperCAmelCase : str = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() __UpperCAmelCase : Optional[int] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Dict = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): __UpperCAmelCase : int = tmp_path / """cache""" __UpperCAmelCase : int = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Any = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = iter_sql_file(__lowerCamelCase ) for rowa, rowa in zip(__lowerCamelCase , __lowerCamelCase ): assert rowa == rowa @require_sqlalchemy def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): __UpperCAmelCase : Union[str, Any] = tmp_path / """cache""" __UpperCAmelCase : Optional[int] = os.path.join(__lowerCamelCase , """tmp.sql""" ) __UpperCAmelCase : Optional[int] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=__lowerCamelCase ).read() with pytest.raises(__lowerCamelCase ): SqlDatasetWriter(__lowerCamelCase , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
63
0
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=lowercase__ ) class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' A : str = field(default='question-answering-extractive' , metadata={'include_in_asdict_even_if_is_default': True} ) A : ClassVar[Features] = Features({'question': Value('string' ), 'context': Value('string' )} ) A : ClassVar[Features] = Features( { 'answers': Sequence( { 'text': Value('string' ), 'answer_start': Value('int32' ), } ) } ) A : str = "question" A : str = "context" A : str = "answers" @property def _lowerCAmelCase ( self ) -> Dict[str, str]: return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
568
from __future__ import annotations a : Optional[Any] = [True] * 1_000_001 a : Union[str, Any] = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): a : Optional[Any] = False i += 1 def lowerCamelCase__ ( __lowerCamelCase : int ): return seive[n] def lowerCamelCase__ ( __lowerCamelCase : int ): return any(digit in """02468""" for digit in str(__lowerCamelCase ) ) def lowerCamelCase__ ( __lowerCamelCase : int = 1000000 ): __UpperCAmelCase : Optional[Any] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__lowerCamelCase ) and not contains_an_even_digit(__lowerCamelCase ): __UpperCAmelCase : Tuple = str(__lowerCamelCase ) __UpperCAmelCase : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__lowerCamelCase ) )] if all(is_prime(__lowerCamelCase ) for i in list_nums ): result.append(__lowerCamelCase ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"""{len(find_circular_primes()) = }""")
63
0