code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: if isinstance(lowerCAmelCase_ , torch.Tensor ): return image elif isinstance(lowerCAmelCase_ , PIL.Image.Image ): lowerCAmelCase_ : Optional[int] = [image] if isinstance(image[0] , PIL.Image.Image ): lowerCAmelCase_ : List[str] = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] lowerCAmelCase_ : str = np.concatenate(lowerCAmelCase_ , axis=0 ) lowerCAmelCase_ : Union[str, Any] = np.array(lowerCAmelCase_ ).astype(np.floataa ) / 255.0 lowerCAmelCase_ : List[Any] = image.transpose(0 , 3 , 1 , 2 ) lowerCAmelCase_ : int = 2.0 * image - 1.0 lowerCAmelCase_ : Optional[int] = torch.from_numpy(lowerCAmelCase_ ) elif isinstance(image[0] , torch.Tensor ): lowerCAmelCase_ : int = torch.cat(lowerCAmelCase_ , dim=0 ) return image def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=0.9995 )-> List[Any]: if not isinstance(lowerCAmelCase_ , np.ndarray ): lowerCAmelCase_ : Tuple = True lowerCAmelCase_ : int = va.device lowerCAmelCase_ : List[str] = va.cpu().numpy() lowerCAmelCase_ : List[str] = va.cpu().numpy() lowerCAmelCase_ : List[Any] = np.sum(va * va / (np.linalg.norm(lowerCAmelCase_ ) * np.linalg.norm(lowerCAmelCase_ )) ) if np.abs(lowerCAmelCase_ ) > DOT_THRESHOLD: lowerCAmelCase_ : List[Any] = (1 - t) * va + t * va else: lowerCAmelCase_ : str = np.arccos(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = np.sin(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = theta_a * t lowerCAmelCase_ : Dict = np.sin(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = np.sin(theta_a - theta_t ) / sin_theta_a lowerCAmelCase_ : Optional[int] = sin_theta_t / sin_theta_a lowerCAmelCase_ : Any = sa * va + sa * va if inputs_are_torch: lowerCAmelCase_ : List[Any] = torch.from_numpy(lowerCAmelCase_ ).to(lowerCAmelCase_ ) return va def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: lowerCAmelCase_ : List[Any] = F.normalize(lowerCAmelCase_ , dim=-1 ) lowerCAmelCase_ : List[Any] = F.normalize(lowerCAmelCase_ , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: for param in model.parameters(): lowerCAmelCase_ : List[Any] = value class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , ) -> Union[str, Any]: super().__init__() self.register_modules( vae=__lowercase , text_encoder=__lowercase , clip_model=__lowercase , tokenizer=__lowercase , unet=__lowercase , scheduler=__lowercase , feature_extractor=__lowercase , coca_model=__lowercase , coca_tokenizer=__lowercase , coca_transform=__lowercase , ) lowerCAmelCase_ : Any = ( feature_extractor.size if isinstance(feature_extractor.size , __lowercase ) else feature_extractor.size['''shortest_edge'''] ) lowerCAmelCase_ : Tuple = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , __lowercase ) set_requires_grad(self.clip_model , __lowercase ) def lowercase_ ( self , __lowercase = "auto" ) -> List[str]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCAmelCase_ : Dict = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__lowercase ) def lowercase_ ( self ) -> Dict: self.enable_attention_slicing(__lowercase ) def lowercase_ ( self ) -> Union[str, Any]: set_requires_grad(self.vae , __lowercase ) def lowercase_ ( self ) -> str: set_requires_grad(self.vae , __lowercase ) def lowercase_ ( self ) -> Union[str, Any]: set_requires_grad(self.unet , __lowercase ) def lowercase_ ( self ) -> Dict: set_requires_grad(self.unet , __lowercase ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Tuple: # get the original timestep using init_timestep lowerCAmelCase_ : Optional[Any] = min(int(num_inference_steps * strength ) , __lowercase ) lowerCAmelCase_ : Dict = max(num_inference_steps - init_timestep , 0 ) lowerCAmelCase_ : int = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None ) -> Tuple: if not isinstance(__lowercase , torch.Tensor ): raise ValueError(f"""`image` has to be of type `torch.Tensor` but is {type(__lowercase )}""" ) lowerCAmelCase_ : List[str] = image.to(device=__lowercase , dtype=__lowercase ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Tuple = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(__lowercase ) ] lowerCAmelCase_ : List[str] = torch.cat(__lowercase , dim=0 ) else: lowerCAmelCase_ : List[str] = self.vae.encode(__lowercase ).latent_dist.sample(__lowercase ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCAmelCase_ : int = 0.1_82_15 * init_latents lowerCAmelCase_ : Union[str, Any] = init_latents.repeat_interleave(__lowercase , dim=0 ) lowerCAmelCase_ : Optional[Any] = randn_tensor(init_latents.shape , generator=__lowercase , device=__lowercase , dtype=__lowercase ) # get latents lowerCAmelCase_ : Dict = self.scheduler.add_noise(__lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = init_latents return latents def lowercase_ ( self , __lowercase ) -> str: lowerCAmelCase_ : int = self.coca_transform(__lowercase ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): lowerCAmelCase_ : Tuple = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) lowerCAmelCase_ : int = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split('''<end_of_text>''' )[0].replace('''<start_of_text>''' , '''''' ).rstrip(''' .,''' ) def lowercase_ ( self , __lowercase , __lowercase ) -> Tuple: lowerCAmelCase_ : Tuple = self.feature_extractor.preprocess(__lowercase ) lowerCAmelCase_ : Tuple = torch.from_numpy(clip_image_input['''pixel_values'''][0] ).unsqueeze(0 ).to(self.device ).half() lowerCAmelCase_ : Dict = self.clip_model.get_image_features(__lowercase ) lowerCAmelCase_ : str = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__lowercase ) lowerCAmelCase_ : int = image_embeddings_clip.repeat_interleave(__lowercase , dim=0 ) return image_embeddings_clip @torch.enable_grad() def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> List[Any]: lowerCAmelCase_ : int = latents.detach().requires_grad_() lowerCAmelCase_ : List[str] = self.scheduler.scale_model_input(__lowercase , __lowercase ) # predict the noise residual lowerCAmelCase_ : List[Any] = self.unet(__lowercase , __lowercase , encoder_hidden_states=__lowercase ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): lowerCAmelCase_ : str = self.scheduler.alphas_cumprod[timestep] lowerCAmelCase_ : Union[str, Any] = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowerCAmelCase_ : Tuple = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 lowerCAmelCase_ : Any = torch.sqrt(__lowercase ) lowerCAmelCase_ : List[str] = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , __lowercase ): lowerCAmelCase_ : Optional[int] = self.scheduler.sigmas[index] lowerCAmelCase_ : str = latents - sigma * noise_pred else: raise ValueError(f"""scheduler type {type(self.scheduler )} not supported""" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCAmelCase_ : Dict = 1 / 0.1_82_15 * sample lowerCAmelCase_ : Any = self.vae.decode(__lowercase ).sample lowerCAmelCase_ : Optional[Any] = (image / 2 + 0.5).clamp(0 , 1 ) lowerCAmelCase_ : List[Any] = transforms.Resize(self.feature_extractor_size )(__lowercase ) lowerCAmelCase_ : List[str] = self.normalize(__lowercase ).to(latents.dtype ) lowerCAmelCase_ : Optional[Any] = self.clip_model.get_image_features(__lowercase ) lowerCAmelCase_ : Optional[Any] = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__lowercase ) lowerCAmelCase_ : int = spherical_dist_loss(__lowercase , __lowercase ).mean() * clip_guidance_scale lowerCAmelCase_ : int = -torch.autograd.grad(__lowercase , __lowercase )[0] if isinstance(self.scheduler , __lowercase ): lowerCAmelCase_ : Union[str, Any] = latents.detach() + grads * (sigma**2) lowerCAmelCase_ : str = noise_pred_original else: lowerCAmelCase_ : Any = noise_pred_original - torch.sqrt(__lowercase ) * grads return noise_pred, latents @torch.no_grad() def __call__( self , __lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = 5_1_2 , __lowercase = 5_1_2 , __lowercase = 0.6 , __lowercase = 5_0 , __lowercase = 7.5 , __lowercase = 1 , __lowercase = 0.0 , __lowercase = 1_0_0 , __lowercase = None , __lowercase = "pil" , __lowercase = True , __lowercase = 0.8 , __lowercase = 0.1 , __lowercase = 0.1 , ) -> Dict: if isinstance(__lowercase , __lowercase ) and len(__lowercase ) != batch_size: raise ValueError(f"""You have passed {batch_size} batch_size, but only {len(__lowercase )} generators.""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if isinstance(__lowercase , torch.Generator ) and batch_size > 1: lowerCAmelCase_ : Optional[int] = [generator] + [None] * (batch_size - 1) lowerCAmelCase_ : List[str] = [ ('''model''', self.coca_model is None), ('''tokenizer''', self.coca_tokenizer is None), ('''transform''', self.coca_transform is None), ] lowerCAmelCase_ : Optional[int] = [x[0] for x in coca_is_none if x[1]] lowerCAmelCase_ : Union[str, Any] = ''', '''.join(__lowercase ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(__lowercase ): raise ValueError( f"""Content prompt is None and CoCa [{coca_is_none_str}] is None.""" f"""Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) lowerCAmelCase_ : Tuple = self.get_image_description(__lowercase ) if style_prompt is None: if len(__lowercase ): raise ValueError( f"""Style prompt is None and CoCa [{coca_is_none_str}] is None.""" f""" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) lowerCAmelCase_ : Tuple = self.get_image_description(__lowercase ) # get prompt text embeddings for content and style lowerCAmelCase_ : List[Any] = self.tokenizer( __lowercase , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=__lowercase , return_tensors='''pt''' , ) lowerCAmelCase_ : Optional[int] = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] lowerCAmelCase_ : Optional[Any] = self.tokenizer( __lowercase , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=__lowercase , return_tensors='''pt''' , ) lowerCAmelCase_ : Optional[int] = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] lowerCAmelCase_ : int = slerp(__lowercase , __lowercase , __lowercase ) # duplicate text embeddings for each generation per prompt lowerCAmelCase_ : Optional[Any] = text_embeddings.repeat_interleave(__lowercase , dim=0 ) # set timesteps lowerCAmelCase_ : List[str] = '''offset''' in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) lowerCAmelCase_ : str = {} if accepts_offset: lowerCAmelCase_ : Any = 1 self.scheduler.set_timesteps(__lowercase , **__lowercase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = self.get_timesteps(__lowercase , __lowercase , self.device ) lowerCAmelCase_ : List[str] = timesteps[:1].repeat(__lowercase ) # Preprocess image lowerCAmelCase_ : Optional[int] = preprocess(__lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = self.prepare_latents( __lowercase , __lowercase , __lowercase , text_embeddings.dtype , self.device , __lowercase ) lowerCAmelCase_ : Dict = preprocess(__lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Any = self.prepare_latents( __lowercase , __lowercase , __lowercase , text_embeddings.dtype , self.device , __lowercase ) lowerCAmelCase_ : Any = slerp(__lowercase , __lowercase , __lowercase ) if clip_guidance_scale > 0: lowerCAmelCase_ : Dict = self.get_clip_image_embeddings(__lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.get_clip_image_embeddings(__lowercase , __lowercase ) lowerCAmelCase_ : str = slerp( __lowercase , __lowercase , __lowercase ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. lowerCAmelCase_ : Union[str, Any] = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: lowerCAmelCase_ : Dict = content_text_input.input_ids.shape[-1] lowerCAmelCase_ : Union[str, Any] = self.tokenizer([''''''] , padding='''max_length''' , max_length=__lowercase , return_tensors='''pt''' ) lowerCAmelCase_ : int = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt lowerCAmelCase_ : Union[str, Any] = uncond_embeddings.repeat_interleave(__lowercase , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCAmelCase_ : Dict = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. lowerCAmelCase_ : Optional[Any] = (batch_size, self.unet.config.in_channels, height // 8, width // 8) lowerCAmelCase_ : Any = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps lowerCAmelCase_ : Union[str, Any] = torch.randn(__lowercase , generator=__lowercase , device='''cpu''' , dtype=__lowercase ).to( self.device ) else: lowerCAmelCase_ : List[str] = torch.randn(__lowercase , generator=__lowercase , device=self.device , dtype=__lowercase ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) lowerCAmelCase_ : str = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase_ : Optional[int] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase_ : int = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase_ : int = {} if accepts_eta: lowerCAmelCase_ : Optional[int] = eta # check if the scheduler accepts generator lowerCAmelCase_ : Dict = '''generator''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: lowerCAmelCase_ : Union[str, Any] = generator with self.progress_bar(total=__lowercase ): for i, t in enumerate(__lowercase ): # expand the latents if we are doing classifier free guidance lowerCAmelCase_ : Dict = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCAmelCase_ : Optional[Any] = self.scheduler.scale_model_input(__lowercase , __lowercase ) # predict the noise residual lowerCAmelCase_ : List[str] = self.unet(__lowercase , __lowercase , encoder_hidden_states=__lowercase ).sample # perform classifier free guidance if do_classifier_free_guidance: lowerCAmelCase_ , lowerCAmelCase_ : Any = noise_pred.chunk(2 ) lowerCAmelCase_ : Optional[int] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: lowerCAmelCase_ : Union[str, Any] = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = self.cond_fn( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase_ : Optional[int] = self.scheduler.step(__lowercase , __lowercase , __lowercase , **__lowercase ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCAmelCase_ : List[str] = 1 / 0.1_82_15 * latents lowerCAmelCase_ : str = self.vae.decode(__lowercase ).sample lowerCAmelCase_ : Optional[int] = (image / 2 + 0.5).clamp(0 , 1 ) lowerCAmelCase_ : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase_ : List[Any] = self.numpy_to_pil(__lowercase ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=__lowercase , nsfw_content_detected=__lowercase )
262
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
1
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : torch.FloatTensor class snake_case__( UpperCAmelCase__, UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , __lowercase = 3_2 , __lowercase = 6_4 , __lowercase = 2_0 , __lowercase = 7_6_8 , __lowercase=7_7 , __lowercase=4 , __lowercase = 0.0 , __lowercase = "silu" , __lowercase = None , __lowercase = None , __lowercase = "linear" , __lowercase = "prd" , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Dict: super().__init__() lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Any = attention_head_dim lowerCAmelCase_ : List[str] = num_attention_heads * attention_head_dim lowerCAmelCase_ : List[str] = additional_embeddings lowerCAmelCase_ : str = time_embed_dim or inner_dim lowerCAmelCase_ : int = embedding_proj_dim or embedding_dim lowerCAmelCase_ : int = clip_embed_dim or embedding_dim lowerCAmelCase_ : Dict = Timesteps(__lowercase , __lowercase , 0 ) lowerCAmelCase_ : Optional[int] = TimestepEmbedding(__lowercase , __lowercase , out_dim=__lowercase , act_fn=__lowercase ) lowerCAmelCase_ : int = nn.Linear(__lowercase , __lowercase ) if embedding_proj_norm_type is None: lowerCAmelCase_ : Optional[int] = None elif embedding_proj_norm_type == "layer": lowerCAmelCase_ : Optional[Any] = nn.LayerNorm(__lowercase ) else: raise ValueError(f"""unsupported embedding_proj_norm_type: {embedding_proj_norm_type}""" ) lowerCAmelCase_ : List[Any] = nn.Linear(__lowercase , __lowercase ) if encoder_hid_proj_type is None: lowerCAmelCase_ : List[Any] = None elif encoder_hid_proj_type == "linear": lowerCAmelCase_ : Tuple = nn.Linear(__lowercase , __lowercase ) else: raise ValueError(f"""unsupported encoder_hid_proj_type: {encoder_hid_proj_type}""" ) lowerCAmelCase_ : Any = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , __lowercase ) ) if added_emb_type == "prd": lowerCAmelCase_ : List[Any] = nn.Parameter(torch.zeros(1 , 1 , __lowercase ) ) elif added_emb_type is None: lowerCAmelCase_ : Union[str, Any] = None else: raise ValueError( f"""`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`.""" ) lowerCAmelCase_ : List[str] = nn.ModuleList( [ BasicTransformerBlock( __lowercase , __lowercase , __lowercase , dropout=__lowercase , activation_fn='''gelu''' , attention_bias=__lowercase , ) for d in range(__lowercase ) ] ) if norm_in_type == "layer": lowerCAmelCase_ : Any = nn.LayerNorm(__lowercase ) elif norm_in_type is None: lowerCAmelCase_ : List[str] = None else: raise ValueError(f"""Unsupported norm_in_type: {norm_in_type}.""" ) lowerCAmelCase_ : List[str] = nn.LayerNorm(__lowercase ) lowerCAmelCase_ : Any = nn.Linear(__lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -1_00_00.0 ) causal_attention_mask.triu_(1 ) lowerCAmelCase_ : Union[str, Any] = causal_attention_mask[None, ...] self.register_buffer('''causal_attention_mask''' , __lowercase , persistent=__lowercase ) lowerCAmelCase_ : int = nn.Parameter(torch.zeros(1 , __lowercase ) ) lowerCAmelCase_ : Any = nn.Parameter(torch.zeros(1 , __lowercase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def lowercase_ ( self ) -> Dict[str, AttentionProcessor]: lowerCAmelCase_ : Union[str, Any] = {} def fn_recursive_add_processors(__lowercase , __lowercase , __lowercase ): if hasattr(__lowercase , '''set_processor''' ): lowerCAmelCase_ : Union[str, Any] = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f"""{name}.{sub_name}""" , __lowercase , __lowercase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(__lowercase , __lowercase , __lowercase ) return processors def lowercase_ ( self , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Union[str, Any] = len(self.attn_processors.keys() ) if isinstance(__lowercase , __lowercase ) and len(__lowercase ) != count: raise ValueError( f"""A dict of processors was passed, but the number of processors {len(__lowercase )} does not match the""" f""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""" ) def fn_recursive_attn_processor(__lowercase , __lowercase , __lowercase ): if hasattr(__lowercase , '''set_processor''' ): if not isinstance(__lowercase , __lowercase ): module.set_processor(__lowercase ) else: module.set_processor(processor.pop(f"""{name}.processor""" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f"""{name}.{sub_name}""" , __lowercase , __lowercase ) for name, module in self.named_children(): fn_recursive_attn_processor(__lowercase , __lowercase , __lowercase ) def lowercase_ ( self ) -> Union[str, Any]: self.set_attn_processor(AttnProcessor() ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = True , ) -> Tuple: lowerCAmelCase_ : Dict = hidden_states.shape[0] lowerCAmelCase_ : int = timestep if not torch.is_tensor(__lowercase ): lowerCAmelCase_ : List[str] = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(__lowercase ) and len(timesteps.shape ) == 0: lowerCAmelCase_ : Tuple = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowerCAmelCase_ : str = timesteps * torch.ones(__lowercase , dtype=timesteps.dtype , device=timesteps.device ) lowerCAmelCase_ : List[str] = self.time_proj(__lowercase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. lowerCAmelCase_ : Optional[Any] = timesteps_projected.to(dtype=self.dtype ) lowerCAmelCase_ : Union[str, Any] = self.time_embedding(__lowercase ) if self.embedding_proj_norm is not None: lowerCAmelCase_ : int = self.embedding_proj_norm(__lowercase ) lowerCAmelCase_ : List[str] = self.embedding_proj(__lowercase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: lowerCAmelCase_ : Dict = self.encoder_hidden_states_proj(__lowercase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError('''`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set''' ) lowerCAmelCase_ : int = self.proj_in(__lowercase ) lowerCAmelCase_ : Any = self.positional_embedding.to(hidden_states.dtype ) lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = 0 if encoder_hidden_states is not None: additional_embeds.append(__lowercase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: lowerCAmelCase_ : Optional[int] = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: lowerCAmelCase_ : Optional[int] = hidden_states[:, None, :] lowerCAmelCase_ : Tuple = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: lowerCAmelCase_ : Tuple = self.prd_embedding.to(hidden_states.dtype ).expand(__lowercase , -1 , -1 ) additional_embeds.append(__lowercase ) lowerCAmelCase_ : str = torch.cat( __lowercase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens lowerCAmelCase_ : Union[str, Any] = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: lowerCAmelCase_ : Dict = F.pad( __lowercase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) lowerCAmelCase_ : List[Any] = hidden_states + positional_embeddings if attention_mask is not None: lowerCAmelCase_ : Optional[int] = (1 - attention_mask.to(hidden_states.dtype )) * -1_00_00.0 lowerCAmelCase_ : List[Any] = F.pad(__lowercase , (0, self.additional_embeddings) , value=0.0 ) lowerCAmelCase_ : List[Any] = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) lowerCAmelCase_ : Tuple = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: lowerCAmelCase_ : Any = self.norm_in(__lowercase ) for block in self.transformer_blocks: lowerCAmelCase_ : Dict = block(__lowercase , attention_mask=__lowercase ) lowerCAmelCase_ : Optional[Any] = self.norm_out(__lowercase ) if self.prd_embedding is not None: lowerCAmelCase_ : Any = hidden_states[:, -1] else: lowerCAmelCase_ : Dict = hidden_states[:, additional_embeddings_len:] lowerCAmelCase_ : Dict = self.proj_to_clip_embeddings(__lowercase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=__lowercase ) def lowercase_ ( self , __lowercase ) -> Optional[Any]: lowerCAmelCase_ : List[str] = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
262
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
1
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _UpperCAmelCase : Any ="""\ Text data. Second line of data.""" _UpperCAmelCase : Union[str, Any] ="""file""" @pytest.fixture(scope='''session''' ) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: lowerCAmelCase_ : List[Any] = tmp_path_factory.mktemp('''data''' ) / (FILE_PATH + '''.zstd''') lowerCAmelCase_ : str = bytes(lowerCAmelCase_ , '''utf-8''' ) with zstd.open(lowerCAmelCase_ , '''wb''' ) as f: f.write(lowerCAmelCase_ ) return path @pytest.fixture def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: with open(os.path.join(tmpfs.local_root_dir , lowerCAmelCase_ ) , '''w''' ) as f: f.write(lowerCAmelCase_ ) return FILE_PATH @pytest.mark.parametrize('''compression_format''' , ['''gzip''', '''xz''', '''zstd'''] ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Any: lowerCAmelCase_ : str = {'''gzip''': gz_file, '''xz''': xz_file, '''zstd''': zstd_path} lowerCAmelCase_ : Union[str, Any] = input_paths[compression_format] lowerCAmelCase_ : List[str] = tmp_path / '''cache''' lowerCAmelCase_ : str = DownloadConfig(cache_dir=lowerCAmelCase_ , extract_compressed_file=lowerCAmelCase_ ) lowerCAmelCase_ : Dict = cached_path(lowerCAmelCase_ , download_config=lowerCAmelCase_ ) with open(lowerCAmelCase_ ) as f: lowerCAmelCase_ : List[str] = f.read() with open(lowerCAmelCase_ ) as f: lowerCAmelCase_ : Optional[int] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('''default_extracted''' , [True, False] ) @pytest.mark.parametrize('''default_cache_dir''' , [True, False] ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[Any]: lowerCAmelCase_ : List[Any] = '''custom_cache''' lowerCAmelCase_ : Dict = '''custom_extracted_dir''' lowerCAmelCase_ : Union[str, Any] = tmp_path / '''custom_extracted_path''' if default_extracted: lowerCAmelCase_ : Tuple = ('''downloads''' if default_cache_dir else custom_cache_dir, '''extracted''') else: monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_DIR''' , lowerCAmelCase_ ) monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(lowerCAmelCase_ ) ) lowerCAmelCase_ : List[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowerCAmelCase_ : str = xz_file lowerCAmelCase_ : Union[str, Any] = ( DownloadConfig(extract_compressed_file=lowerCAmelCase_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=lowerCAmelCase_ ) ) lowerCAmelCase_ : Optional[Any] = cached_path(lowerCAmelCase_ , download_config=lowerCAmelCase_ ) assert Path(lowerCAmelCase_ ).parent.parts[-2:] == expected def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: # absolute path lowerCAmelCase_ : Optional[Any] = str(Path(lowerCAmelCase_ ).resolve() ) assert cached_path(lowerCAmelCase_ ) == text_file # relative path lowerCAmelCase_ : Tuple = str(Path(lowerCAmelCase_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowerCAmelCase_ ) == text_file def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: # absolute path lowerCAmelCase_ : List[Any] = str(tmp_path.resolve() / '''__missing_file__.txt''' ) with pytest.raises(lowerCAmelCase_ ): cached_path(lowerCAmelCase_ ) # relative path lowerCAmelCase_ : Dict = '''./__missing_file__.txt''' with pytest.raises(lowerCAmelCase_ ): cached_path(lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: lowerCAmelCase_ : Dict = get_from_cache(f"""tmp://{tmpfs_file}""" ) with open(lowerCAmelCase_ ) as f: lowerCAmelCase_ : Tuple = f.read() assert output_file_content == FILE_CONTENT @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def lowerCAmelCase ( )-> Dict: with pytest.raises(lowerCAmelCase_ ): cached_path('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : Tuple = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): http_get('''https://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): http_head('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : Any = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): ftp_get('''ftp://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): ftp_head('''ftp://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> Dict: lowerCAmelCase_ : List[str] = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): fsspec_get('''s3://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): fsspec_head('''s3://huggingface.co''' )
262
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Union[str, Any] ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Union[str, Any] =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _UpperCAmelCase : Dict =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _UpperCAmelCase : Union[str, Any] =HfArgumentParser(InitializationArguments) _UpperCAmelCase : Tuple =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization _UpperCAmelCase : str =AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks _UpperCAmelCase : Optional[int] ={ """vocab_size""": len(tokenizer), """scale_attn_by_inverse_layer_idx""": True, """reorder_and_upcast_attn""": True, } # Load model config (GPT-2 large in this case) _UpperCAmelCase : Optional[Any] =AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config _UpperCAmelCase : Tuple =AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
262
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
1
import flax.linen as nn import jax import jax.numpy as jnp class snake_case__( nn.Module ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa def lowercase_ ( self ) -> Any: lowerCAmelCase_ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , __lowercase ) -> Dict: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = hidden_states.shape lowerCAmelCase_ : List[str] = jax.image.resize( __lowercase , shape=(batch, height * 2, width * 2, channels) , method='''nearest''' , ) lowerCAmelCase_ : List[Any] = self.conv(__lowercase ) return hidden_states class snake_case__( nn.Module ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : List[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , __lowercase ) -> Any: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCAmelCase_ : Optional[int] = self.conv(__lowercase ) return hidden_states class snake_case__( nn.Module ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int SCREAMING_SNAKE_CASE__ : int = None SCREAMING_SNAKE_CASE__ : float = 0.0 SCREAMING_SNAKE_CASE__ : bool = None SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : int = self.in_channels if self.out_channels is None else self.out_channels lowerCAmelCase_ : Optional[Any] = nn.GroupNorm(num_groups=3_2 , epsilon=1e-5 ) lowerCAmelCase_ : Dict = nn.Conv( __lowercase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCAmelCase_ : Any = nn.Dense(__lowercase , dtype=self.dtype ) lowerCAmelCase_ : Union[str, Any] = nn.GroupNorm(num_groups=3_2 , epsilon=1e-5 ) lowerCAmelCase_ : List[str] = nn.Dropout(self.dropout_prob ) lowerCAmelCase_ : Any = nn.Conv( __lowercase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCAmelCase_ : List[str] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCAmelCase_ : Optional[int] = None if use_nin_shortcut: lowerCAmelCase_ : Any = nn.Conv( __lowercase , kernel_size=(1, 1) , strides=(1, 1) , padding='''VALID''' , dtype=self.dtype , ) def __call__( self , __lowercase , __lowercase , __lowercase=True ) -> List[Any]: lowerCAmelCase_ : int = hidden_states lowerCAmelCase_ : int = self.norma(__lowercase ) lowerCAmelCase_ : Any = nn.swish(__lowercase ) lowerCAmelCase_ : Optional[int] = self.conva(__lowercase ) lowerCAmelCase_ : Optional[int] = self.time_emb_proj(nn.swish(__lowercase ) ) lowerCAmelCase_ : Tuple = jnp.expand_dims(jnp.expand_dims(__lowercase , 1 ) , 1 ) lowerCAmelCase_ : List[str] = hidden_states + temb lowerCAmelCase_ : List[Any] = self.norma(__lowercase ) lowerCAmelCase_ : int = nn.swish(__lowercase ) lowerCAmelCase_ : str = self.dropout(__lowercase , __lowercase ) lowerCAmelCase_ : Tuple = self.conva(__lowercase ) if self.conv_shortcut is not None: lowerCAmelCase_ : int = self.conv_shortcut(__lowercase ) return hidden_states + residual
262
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
1
class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : int = len(__lowercase ) lowerCAmelCase_ : Dict = [0] * len_array if len_array > 0: lowerCAmelCase_ : List[Any] = array[0] for i in range(1 , __lowercase ): lowerCAmelCase_ : int = self.prefix_sum[i - 1] + array[i] def lowercase_ ( self , __lowercase , __lowercase ) -> int: if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowercase_ ( self , __lowercase ) -> bool: lowerCAmelCase_ : Optional[Any] = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(__lowercase ) return False if __name__ == "__main__": import doctest doctest.testmod()
262
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
1
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class snake_case__( UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = IFInpaintingSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def lowercase_ ( self ) -> Optional[int]: return self._get_superresolution_dummy_components() def lowercase_ ( self , __lowercase , __lowercase=0 ) -> Union[str, Any]: if str(__lowercase ).startswith('''mps''' ): lowerCAmelCase_ : Union[str, Any] = torch.manual_seed(__lowercase ) else: lowerCAmelCase_ : int = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) lowerCAmelCase_ : Tuple = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(__lowercase ) ).to(__lowercase ) lowerCAmelCase_ : Optional[int] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__lowercase ) ).to(__lowercase ) lowerCAmelCase_ : Union[str, Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__lowercase ) ).to(__lowercase ) lowerCAmelCase_ : List[Any] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowercase_ ( self ) -> Tuple: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowercase_ ( self ) -> List[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def lowercase_ ( self ) -> str: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowercase_ ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowercase_ ( self ) -> List[str]: self._test_save_load_local() def lowercase_ ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
262
from __future__ import annotations import math class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : str = size # approximate the overall size of segment tree with given value lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] lowerCAmelCase_ : Optional[int] = [0 for i in range(0 , 4 * size )] # flag for lazy update def lowercase_ ( self , __lowercase ) -> int: return idx * 2 def lowercase_ ( self , __lowercase ) -> int: return idx * 2 + 1 def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> None: if left_element == right_element: lowerCAmelCase_ : Tuple = a[left_element - 1] else: lowerCAmelCase_ : int = (left_element + right_element) // 2 self.build(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase ) self.build(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase ) lowerCAmelCase_ : Any = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> bool: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Union[str, Any] = False if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Any = self.lazy[idx] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCAmelCase_ : Dict = val if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = val lowerCAmelCase_ : List[Any] = val lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : List[str] = True return True lowerCAmelCase_ : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) self.update(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) return True def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> int | float: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Optional[Any] = False if left_element != right_element: lowerCAmelCase_ : List[Any] = self.lazy[idx] lowerCAmelCase_ : Dict = self.lazy[idx] lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCAmelCase_ : List[Any] = (left_element + right_element) // 2 lowerCAmelCase_ : Tuple = self.query(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.query(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase ) return max(__lowercase , __lowercase ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , __lowercase , __lowercase ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _UpperCAmelCase : str =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _UpperCAmelCase : List[str] =15 _UpperCAmelCase : Any =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
262
1
import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class snake_case__: '''simple docstring''' def __init__( self , __lowercase , __lowercase=1_3 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=False , __lowercase=True , __lowercase=9_9 , __lowercase=6_4 , __lowercase=5 , __lowercase=4 , __lowercase=6_4 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_1_2 , __lowercase=1_6 , __lowercase=2 , __lowercase=0.02 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> Optional[int]: lowerCAmelCase_ : Optional[int] = parent lowerCAmelCase_ : int = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : str = is_training lowerCAmelCase_ : Optional[Any] = use_input_mask lowerCAmelCase_ : Union[str, Any] = use_token_type_ids lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Optional[int] = hidden_size lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Optional[Any] = num_attention_heads lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Optional[int] = hidden_act lowerCAmelCase_ : int = hidden_dropout_prob lowerCAmelCase_ : Union[str, Any] = attention_probs_dropout_prob lowerCAmelCase_ : Optional[int] = max_position_embeddings lowerCAmelCase_ : Optional[Any] = type_vocab_size lowerCAmelCase_ : List[Any] = type_sequence_label_size lowerCAmelCase_ : Tuple = initializer_range lowerCAmelCase_ : Optional[int] = num_labels lowerCAmelCase_ : List[Any] = num_choices lowerCAmelCase_ : Union[str, Any] = scope def lowercase_ ( self ) -> Optional[Any]: return MPNetConfig.from_pretrained('''microsoft/mpnet-base''' ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ : Optional[int] = None if self.use_input_mask: lowerCAmelCase_ : str = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : Dict = None lowerCAmelCase_ : Any = None lowerCAmelCase_ : int = None if self.use_labels: lowerCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ : Optional[Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase_ ( self ) -> Tuple: return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : List[str] = MPNetModel(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Dict = model(__lowercase , __lowercase ) lowerCAmelCase_ : str = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Any: lowerCAmelCase_ : Optional[Any] = MPNetForQuestionAnswering(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Optional[Any] = model( __lowercase , attention_mask=__lowercase , start_positions=__lowercase , end_positions=__lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : List[Any] = self.num_labels lowerCAmelCase_ : Any = MPNetForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : int = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : List[str] = self.num_choices lowerCAmelCase_ : str = MPNetForMultipleChoice(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Dict = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : str = model( __lowercase , attention_mask=__lowercase , labels=__lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[Any]: lowerCAmelCase_ : Tuple = self.num_labels lowerCAmelCase_ : List[Any] = MPNetForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Optional[Any] = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() ((lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_)) : List[Any] = config_and_inputs lowerCAmelCase_ : Dict = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class snake_case__( UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ : List[str] = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : List[Any] = False SCREAMING_SNAKE_CASE__ : Dict = True def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Union[str, Any] = MPNetModelTester(self ) lowerCAmelCase_ : int = ConfigTester(self , config_class=__lowercase , hidden_size=3_7 ) def lowercase_ ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def lowercase_ ( self ) -> Any: lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*__lowercase ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*__lowercase ) @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Dict = MPNetModel.from_pretrained('''microsoft/mpnet-base''' ) lowerCAmelCase_ : int = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) lowerCAmelCase_ : int = model(__lowercase )[0] lowerCAmelCase_ : Union[str, Any] = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , __lowercase ) lowerCAmelCase_ : Optional[Any] = torch.tensor( [[[-0.05_50, 0.19_43, -0.07_40], [-0.05_62, 0.22_11, -0.05_79], [-0.04_37, 0.33_37, -0.06_41]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=1e-4 ) )
262
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py _UpperCAmelCase : Optional[int] ="""src/transformers""" _UpperCAmelCase : str ="""docs/source/en""" _UpperCAmelCase : Optional[int] =""".""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : int = f.readlines() # Find the start prompt. lowerCAmelCase_ : List[Any] = 0 while not lines[start_index].startswith(lowerCAmelCase_ ): start_index += 1 start_index += 1 lowerCAmelCase_ : List[str] = start_index while not lines[end_index].startswith(lowerCAmelCase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | _UpperCAmelCase : Optional[Any] ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. _UpperCAmelCase : Optional[int] =re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _UpperCAmelCase : Dict =re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _UpperCAmelCase : Optional[Any] =re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : Optional[int] =direct_transformers_import(TRANSFORMERS_PATH) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : str = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' , lowerCAmelCase_ ) return [m.group(0 ) for m in matches] def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Tuple = 2 if text == '''✅''' or text == '''❌''' else len(lowerCAmelCase_ ) lowerCAmelCase_ : int = (width - text_length) // 2 lowerCAmelCase_ : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def lowerCAmelCase ( )-> str: lowerCAmelCase_ : Any = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowerCAmelCase_ : Dict = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } lowerCAmelCase_ : List[Any] = {name: config.replace('''Config''' , '''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. lowerCAmelCase_ : Tuple = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = None if attr_name.endswith('''Tokenizer''' ): lowerCAmelCase_ : Union[str, Any] = slow_tokenizers lowerCAmelCase_ : List[str] = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): lowerCAmelCase_ : int = fast_tokenizers lowerCAmelCase_ : Union[str, Any] = attr_name[:-13] elif _re_tf_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = tf_models lowerCAmelCase_ : str = _re_tf_models.match(lowerCAmelCase_ ).groups()[0] elif _re_flax_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = flax_models lowerCAmelCase_ : Union[str, Any] = _re_flax_models.match(lowerCAmelCase_ ).groups()[0] elif _re_pt_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Any = pt_models lowerCAmelCase_ : List[Any] = _re_pt_models.match(lowerCAmelCase_ ).groups()[0] if lookup_dict is not None: while len(lowerCAmelCase_ ) > 0: if attr_name in model_name_to_prefix.values(): lowerCAmelCase_ : Union[str, Any] = True break # Try again after removing the last word in the name lowerCAmelCase_ : Any = ''''''.join(camel_case_split(lowerCAmelCase_ )[:-1] ) # Let's build that table! lowerCAmelCase_ : int = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) lowerCAmelCase_ : Tuple = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). lowerCAmelCase_ : Union[str, Any] = [len(lowerCAmelCase_ ) + 2 for c in columns] lowerCAmelCase_ : Optional[Any] = max([len(lowerCAmelCase_ ) for name in model_names] ) + 2 # Build the table per se lowerCAmelCase_ : Dict = '''|''' + '''|'''.join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for c, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" lowerCAmelCase_ : List[str] = {True: '''✅''', False: '''❌'''} for name in model_names: lowerCAmelCase_ : List[Any] = model_name_to_prefix[name] lowerCAmelCase_ : Union[str, Any] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for l, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + "|\n" return table def lowerCAmelCase ( lowerCAmelCase_=False )-> Tuple: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = _find_text_in_file( filename=os.path.join(lowerCAmelCase_ , '''index.md''' ) , start_prompt='''<!--This table is updated automatically from the auto modules''' , end_prompt='''<!-- End table-->''' , ) lowerCAmelCase_ : Tuple = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCAmelCase_ , '''index.md''' ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Tuple =parser.parse_args() check_model_table(args.fix_and_overwrite)
262
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Optional[int] =logging.get_logger(__name__) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_=False )-> List[Any]: lowerCAmelCase_ : Any = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowerCAmelCase_ : Union[str, Any] = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False )-> str: for i in range(config.num_hidden_layers ): if base_model: lowerCAmelCase_ : int = '''''' else: lowerCAmelCase_ : Tuple = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase_ : Optional[int] = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) lowerCAmelCase_ : Optional[int] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase_ : List[str] = in_proj_weight[ : config.hidden_size, : ] lowerCAmelCase_ : List[str] = in_proj_bias[: config.hidden_size] lowerCAmelCase_ : Dict = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase_ : Union[str, Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase_ : Tuple = in_proj_weight[ -config.hidden_size :, : ] lowerCAmelCase_ : Optional[Any] = in_proj_bias[-config.hidden_size :] def lowerCAmelCase ( lowerCAmelCase_ )-> int: lowerCAmelCase_ : str = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : List[str] = dct.pop(lowerCAmelCase_ ) lowerCAmelCase_ : Dict = val def lowerCAmelCase ( )-> List[Any]: lowerCAmelCase_ : Dict = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase_ : Dict = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: lowerCAmelCase_ : Tuple = ViTConfig() lowerCAmelCase_ : Optional[int] = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": lowerCAmelCase_ : str = True lowerCAmelCase_ : str = int(vit_name[-12:-10] ) lowerCAmelCase_ : Tuple = int(vit_name[-9:-6] ) else: lowerCAmelCase_ : Optional[Any] = 1_000 lowerCAmelCase_ : List[Any] = '''huggingface/label-files''' lowerCAmelCase_ : Dict = '''imagenet-1k-id2label.json''' lowerCAmelCase_ : Dict = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase_ : Any = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} lowerCAmelCase_ : Any = idalabel lowerCAmelCase_ : List[str] = {v: k for k, v in idalabel.items()} lowerCAmelCase_ : Optional[int] = int(vit_name[-6:-4] ) lowerCAmelCase_ : Dict = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): lowerCAmelCase_ : str = 192 lowerCAmelCase_ : Tuple = 768 lowerCAmelCase_ : List[str] = 12 lowerCAmelCase_ : int = 3 elif vit_name[9:].startswith('''small''' ): lowerCAmelCase_ : Dict = 384 lowerCAmelCase_ : List[str] = 1_536 lowerCAmelCase_ : List[Any] = 12 lowerCAmelCase_ : Union[str, Any] = 6 else: pass else: if vit_name[4:].startswith('''small''' ): lowerCAmelCase_ : Any = 768 lowerCAmelCase_ : str = 2_304 lowerCAmelCase_ : List[str] = 8 lowerCAmelCase_ : List[str] = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): lowerCAmelCase_ : Optional[Any] = 1_024 lowerCAmelCase_ : int = 4_096 lowerCAmelCase_ : List[Any] = 24 lowerCAmelCase_ : Union[str, Any] = 16 elif vit_name[4:].startswith('''huge''' ): lowerCAmelCase_ : int = 1_280 lowerCAmelCase_ : Tuple = 5_120 lowerCAmelCase_ : List[Any] = 32 lowerCAmelCase_ : Optional[Any] = 16 # load original model from timm lowerCAmelCase_ : Optional[Any] = timm.create_model(lowerCAmelCase_ , pretrained=lowerCAmelCase_ ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowerCAmelCase_ : Union[str, Any] = timm_model.state_dict() if base_model: remove_classification_head_(lowerCAmelCase_ ) lowerCAmelCase_ : Any = create_rename_keys(lowerCAmelCase_ , lowerCAmelCase_ ) for src, dest in rename_keys: rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) read_in_q_k_v(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # load HuggingFace model if vit_name[-5:] == "in21k": lowerCAmelCase_ : List[str] = ViTModel(lowerCAmelCase_ ).eval() else: lowerCAmelCase_ : List[Any] = ViTForImageClassification(lowerCAmelCase_ ).eval() model.load_state_dict(lowerCAmelCase_ ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: lowerCAmelCase_ : Tuple = DeiTImageProcessor(size=config.image_size ) else: lowerCAmelCase_ : Optional[Any] = ViTImageProcessor(size=config.image_size ) lowerCAmelCase_ : List[Any] = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowerCAmelCase_ : int = encoding['''pixel_values'''] lowerCAmelCase_ : List[str] = model(lowerCAmelCase_ ) if base_model: lowerCAmelCase_ : int = timm_model.forward_features(lowerCAmelCase_ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(lowerCAmelCase_ , outputs.pooler_output , atol=1e-3 ) else: lowerCAmelCase_ : Optional[Any] = timm_model(lowerCAmelCase_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase_ , outputs.logits , atol=1e-3 ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _UpperCAmelCase : str =argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : Dict =parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
262
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] ={"""vocab_file""": """spiece.model"""} _UpperCAmelCase : List[str] ={ """vocab_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model""", } } _UpperCAmelCase : List[Any] ={ """xlnet-base-cased""": None, """xlnet-large-cased""": None, } # Segments (not really needed) _UpperCAmelCase : Tuple =0 _UpperCAmelCase : Tuple =1 _UpperCAmelCase : int =2 _UpperCAmelCase : List[str] =3 _UpperCAmelCase : int =4 class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = """left""" def __init__( self , __lowercase , __lowercase=False , __lowercase=True , __lowercase=False , __lowercase="<s>" , __lowercase="</s>" , __lowercase="<unk>" , __lowercase="<sep>" , __lowercase="<pad>" , __lowercase="<cls>" , __lowercase="<mask>" , __lowercase=["<eop>", "<eod>"] , __lowercase = None , **__lowercase , ) -> None: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : str = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__lowercase , remove_space=__lowercase , keep_accents=__lowercase , bos_token=__lowercase , eos_token=__lowercase , unk_token=__lowercase , sep_token=__lowercase , pad_token=__lowercase , cls_token=__lowercase , mask_token=__lowercase , additional_special_tokens=__lowercase , sp_model_kwargs=self.sp_model_kwargs , **__lowercase , ) lowerCAmelCase_ : Optional[Any] = 3 lowerCAmelCase_ : List[str] = do_lower_case lowerCAmelCase_ : Optional[int] = remove_space lowerCAmelCase_ : Any = keep_accents lowerCAmelCase_ : Optional[Any] = vocab_file lowerCAmelCase_ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowercase ) @property def lowercase_ ( self ) -> Any: return len(self.sp_model ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Dict = {self.convert_ids_to_tokens(__lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Dict: lowerCAmelCase_ : Dict = self.__dict__.copy() lowerCAmelCase_ : Union[str, Any] = None return state def __setstate__( self , __lowercase ) -> Tuple: lowerCAmelCase_ : Any = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowerCAmelCase_ : int = {} lowerCAmelCase_ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase_ ( self , __lowercase ) -> Optional[Any]: if self.remove_space: lowerCAmelCase_ : Tuple = ''' '''.join(inputs.strip().split() ) else: lowerCAmelCase_ : Tuple = inputs lowerCAmelCase_ : str = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: lowerCAmelCase_ : List[Any] = unicodedata.normalize('''NFKD''' , __lowercase ) lowerCAmelCase_ : Any = ''''''.join([c for c in outputs if not unicodedata.combining(__lowercase )] ) if self.do_lower_case: lowerCAmelCase_ : Dict = outputs.lower() return outputs def lowercase_ ( self , __lowercase ) -> List[str]: lowerCAmelCase_ : Dict = self.preprocess_text(__lowercase ) lowerCAmelCase_ : Optional[Any] = self.sp_model.encode(__lowercase , out_type=__lowercase ) lowerCAmelCase_ : int = [] for piece in pieces: if len(__lowercase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): lowerCAmelCase_ : List[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(__lowercase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCAmelCase_ : Optional[Any] = cur_pieces[1:] else: lowerCAmelCase_ : str = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__lowercase ) else: new_pieces.append(__lowercase ) return new_pieces def lowercase_ ( self , __lowercase ) -> Optional[Any]: return self.sp_model.PieceToId(__lowercase ) def lowercase_ ( self , __lowercase ) -> Optional[int]: return self.sp_model.IdToPiece(__lowercase ) def lowercase_ ( self , __lowercase ) -> Optional[Any]: lowerCAmelCase_ : Dict = ''''''.join(__lowercase ).replace(__lowercase , ''' ''' ).strip() return out_string def lowercase_ ( self , __lowercase , __lowercase = False , __lowercase = None , __lowercase = True , **__lowercase , ) -> str: lowerCAmelCase_ : List[Any] = kwargs.pop('''use_source_tokenizer''' , __lowercase ) lowerCAmelCase_ : Any = self.convert_ids_to_tokens(__lowercase , skip_special_tokens=__lowercase ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Tuple = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__lowercase ) ) lowerCAmelCase_ : int = [] sub_texts.append(__lowercase ) else: current_sub_text.append(__lowercase ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__lowercase ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowerCAmelCase_ : Any = ''''''.join(__lowercase ) lowerCAmelCase_ : Tuple = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCAmelCase_ : Dict = self.clean_up_tokenization(__lowercase ) return clean_text else: return text def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Any = [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowercase_ ( self , __lowercase , __lowercase = None , __lowercase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase ) if token_ids_a is not None: return ([0] * len(__lowercase )) + [1] + ([0] * len(__lowercase )) + [1, 1] return ([0] * len(__lowercase )) + [1, 1] def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Union[str, Any] = [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase_ : List[Any] = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowercase ) elif not os.path.isfile(self.vocab_file ): with open(__lowercase , '''wb''' ) as fi: lowerCAmelCase_ : Dict = self.sp_model.serialized_model_proto() fi.write(__lowercase ) return (out_vocab_file,)
262
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : Tuple =None _UpperCAmelCase : int =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Any ={ """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """facebook/nllb-large-en-ro""": 1024, """facebook/nllb-200-distilled-600M""": 1024, } # fmt: off _UpperCAmelCase : Any =["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : int = NllbTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=False , **__lowercase , ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : int = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : List[Any] = legacy_behaviour super().__init__( vocab_file=__lowercase , tokenizer_file=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , additional_special_tokens=__lowercase , legacy_behaviour=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_file lowerCAmelCase_ : List[Any] = False if not self.vocab_file else True lowerCAmelCase_ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) lowerCAmelCase_ : Optional[Any] = { lang_code: self.convert_tokens_to_ids(__lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase_ : Any = src_lang if src_lang is not None else '''eng_Latn''' lowerCAmelCase_ : str = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase_ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowercase_ ( self ) -> str: return self._src_lang @src_lang.setter def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : int = self(__lowercase , add_special_tokens=__lowercase , return_tensors=__lowercase , **__lowercase ) lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) lowerCAmelCase_ : List[Any] = tgt_lang_id return inputs def lowercase_ ( self , __lowercase , __lowercase = "eng_Latn" , __lowercase = None , __lowercase = "fra_Latn" , **__lowercase , ) -> BatchEncoding: lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : List[str] = tgt_lang return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def lowercase_ ( self ) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[str] = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[str] = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Optional[int] = [self.cur_lang_code] lowerCAmelCase_ : List[Any] = [self.eos_token_id] lowerCAmelCase_ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Any = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Any = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Any = [self.cur_lang_code] lowerCAmelCase_ : Any = [self.eos_token_id] lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ): copyfile(self.vocab_file , __lowercase ) return (out_vocab_file,)
262
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _UpperCAmelCase : Optional[Any] =NewType("""DataClass""", Any) _UpperCAmelCase : Dict =NewType("""DataClassType""", Any) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase ( lowerCAmelCase_ )-> Callable[[str], Any]: lowerCAmelCase_ : str = {str(lowerCAmelCase_ ): choice for choice in choices} return lambda lowerCAmelCase_ : str_to_choice.get(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( *, lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = None , **lowerCAmelCase_ , )-> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase_ : Dict = {} if aliases is not None: lowerCAmelCase_ : str = aliases if help is not None: lowerCAmelCase_ : Tuple = help return dataclasses.field(metadata=lowerCAmelCase_ , default=lowerCAmelCase_ , default_factory=lowerCAmelCase_ , **lowerCAmelCase_ ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Iterable[DataClassType] def __init__( self , __lowercase , **__lowercase ) -> List[str]: # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase_ : Optional[int] = ArgumentDefaultsHelpFormatter super().__init__(**__lowercase ) if dataclasses.is_dataclass(__lowercase ): lowerCAmelCase_ : Union[str, Any] = [dataclass_types] lowerCAmelCase_ : List[Any] = list(__lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowercase ) @staticmethod def lowercase_ ( __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[Any] = f"""--{field.name}""" lowerCAmelCase_ : Tuple = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowercase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase_ : List[str] = kwargs.pop('''aliases''' , [] ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[Any] = [aliases] lowerCAmelCase_ : Any = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(__lowercase , '''UnionType''' ) and isinstance(__lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowercase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(__lowercase ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase_ : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase_ : Dict = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase_ : str = ( field.type.__args__[0] if isinstance(__lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase_ : List[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase_ : Dict = {} if origin_type is Literal or (isinstance(field.type , __lowercase ) and issubclass(field.type , __lowercase )): if origin_type is Literal: lowerCAmelCase_ : Optional[Any] = field.type.__args__ else: lowerCAmelCase_ : int = [x.value for x in field.type] lowerCAmelCase_ : str = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase_ : str = field.default else: lowerCAmelCase_ : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase_ : Tuple = copy(__lowercase ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase_ : Dict = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase_ : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase_ : int = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase_ : List[Any] = True elif isclass(__lowercase ) and issubclass(__lowercase , __lowercase ): lowerCAmelCase_ : Union[str, Any] = field.type.__args__[0] lowerCAmelCase_ : Dict = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : Any = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase_ : Dict = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : List[Any] = field.default_factory() else: lowerCAmelCase_ : int = True parser.add_argument(__lowercase , *__lowercase , **__lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase_ : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: if hasattr(__lowercase , '''_argument_group_name''' ): lowerCAmelCase_ : str = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase_ : Dict = self try: lowerCAmelCase_ : Dict[str, type] = get_type_hints(__lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(__lowercase ): lowerCAmelCase_ : Any = '''.'''.join(map(__lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(__lowercase ): if not field.init: continue lowerCAmelCase_ : Optional[int] = type_hints[field.name] self._parse_dataclass_field(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase=None , __lowercase=False , __lowercase=True , __lowercase=None , __lowercase=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase_ : str = [] if args_filename: args_files.append(Path(__lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase_ : str = ArgumentParser() args_file_parser.add_argument(__lowercase , type=__lowercase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = args_file_parser.parse_known_args(args=__lowercase ) lowerCAmelCase_ : int = vars(__lowercase ).get(args_file_flag.lstrip('''-''' ) , __lowercase ) if cmd_args_file_paths: args_files.extend([Path(__lowercase ) for p in cmd_args_file_paths] ) lowerCAmelCase_ : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.parse_known_args(args=__lowercase ) lowerCAmelCase_ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase_ : str = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : str = {k: v for k, v in vars(__lowercase ).items() if k in keys} for k in keys: delattr(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = dtype(**__lowercase ) outputs.append(__lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : int = set(args.keys() ) lowerCAmelCase_ : str = [] for dtype in self.dataclass_types: lowerCAmelCase_ : int = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : List[str] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase_ : List[str] = dtype(**__lowercase ) outputs.append(__lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(__lowercase )}""" ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: with open(Path(__lowercase ) , encoding='''utf-8''' ) as open_json_file: lowerCAmelCase_ : Dict = json.loads(open_json_file.read() ) lowerCAmelCase_ : str = self.parse_dict(__lowercase , allow_extra_keys=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : Optional[Any] = self.parse_dict(yaml.safe_load(Path(__lowercase ).read_text() ) , allow_extra_keys=__lowercase ) return tuple(__lowercase )
262
1
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets _UpperCAmelCase : Dict ="""\ @inproceedings{popovic-2015-chrf, title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\", author = \"Popovi{\'c}, Maja\", booktitle = \"Proceedings of the Tenth Workshop on Statistical Machine Translation\", month = sep, year = \"2015\", address = \"Lisbon, Portugal\", publisher = \"Association for Computational Linguistics\", url = \"https://aclanthology.org/W15-3049\", doi = \"10.18653/v1/W15-3049\", pages = \"392--395\", } @inproceedings{popovic-2017-chrf, title = \"chr{F}++: words helping character n-grams\", author = \"Popovi{\'c}, Maja\", booktitle = \"Proceedings of the Second Conference on Machine Translation\", month = sep, year = \"2017\", address = \"Copenhagen, Denmark\", publisher = \"Association for Computational Linguistics\", url = \"https://aclanthology.org/W17-4770\", doi = \"10.18653/v1/W17-4770\", pages = \"612--618\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ _UpperCAmelCase : Any ="""\ ChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches, and ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation that is already present in sacrebleu. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information. """ _UpperCAmelCase : int =""" Produces ChrF(++) scores for hypotheses given reference translations. Args: predictions (list of str): The predicted sentences. references (list of list of str): The references. There should be one reference sub-list for each prediction sentence. char_order (int): Character n-gram order. Defaults to `6`. word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`. beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`. lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`. whitespace (bool): If `True`, include whitespaces when extracting character n-grams. eps_smoothing (bool): If `True`, applies epsilon smoothing similar to reference chrF++.py, NLTK and Moses implementations. If `False`, it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`. Returns: 'score' (float): The chrF (chrF++) score, 'char_order' (int): The character n-gram order, 'word_order' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++, 'beta' (int): Determine the importance of recall w.r.t precision Examples: Example 1--a simple example of calculating chrF: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, references=reference) >>> print(results) {'score': 84.64214891738334, 'char_order': 6, 'word_order': 0, 'beta': 2} Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2) >>> print(results) {'score': 82.87263732906315, 'char_order': 6, 'word_order': 2, 'beta': 2} Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2, ... lowercase=True) >>> print(results) {'score': 92.12853119829202, 'char_order': 6, 'word_order': 2, 'beta': 2} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case__( datasets.Metric ): '''simple docstring''' def lowercase_ ( self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse('''1.4.12''' ): raise ImportWarning( '''To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n''' '''You can install it with `pip install "sacrebleu>=1.4.12"`.''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/mjpost/sacreBLEU#chrf--chrf''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=['''https://github.com/mjpost/sacreBLEU#chrf--chrf'''] , reference_urls=[ '''https://github.com/m-popovic/chrF''', ] , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase = CHRF.CHAR_ORDER , __lowercase = CHRF.WORD_ORDER , __lowercase = CHRF.BETA , __lowercase = False , __lowercase = False , __lowercase = False , ) -> Dict: lowerCAmelCase_ : Any = len(references[0] ) if any(len(__lowercase ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) lowerCAmelCase_ : Any = [[refs[i] for refs in references] for i in range(__lowercase )] lowerCAmelCase_ : Dict = CHRF(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Tuple = sb_chrf.corpus_score(__lowercase , __lowercase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
262
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: return EnvironmentCommand() class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @staticmethod def lowercase_ ( __lowercase ) -> List[Any]: lowerCAmelCase_ : List[str] = parser.add_parser('''env''' ) download_parser.set_defaults(func=__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = huggingface_hub.__version__ lowerCAmelCase_ : str = '''not installed''' lowerCAmelCase_ : str = '''NA''' if is_torch_available(): import torch lowerCAmelCase_ : Any = torch.__version__ lowerCAmelCase_ : str = torch.cuda.is_available() lowerCAmelCase_ : List[str] = '''not installed''' if is_transformers_available(): import transformers lowerCAmelCase_ : Any = transformers.__version__ lowerCAmelCase_ : Optional[Any] = '''not installed''' if is_accelerate_available(): import accelerate lowerCAmelCase_ : List[Any] = accelerate.__version__ lowerCAmelCase_ : List[str] = '''not installed''' if is_xformers_available(): import xformers lowerCAmelCase_ : Optional[Any] = xformers.__version__ lowerCAmelCase_ : int = { '''`diffusers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Huggingface_hub version''': hub_version, '''Transformers version''': transformers_version, '''Accelerate version''': accelerate_version, '''xFormers version''': xformers_version, '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def lowercase_ ( __lowercase ) -> str: return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
262
1
import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer _UpperCAmelCase : Dict =logging.get_logger(__name__) _UpperCAmelCase : Union[str, Any] ={"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[str] ={ """vocab_file""": { """Salesforce/codegen-350M-mono""": """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json""", }, """merges_file""": { """Salesforce/codegen-350M-mono""": """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt""", }, """tokenizer_file""": { """Salesforce/codegen-350M-mono""": ( """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """Salesforce/codegen-350M-mono""": 2048, } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : int = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : int = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : List[Any] = CodeGenTokenizer def __init__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase="<|endoftext|>" , __lowercase="<|endoftext|>" , __lowercase="<|endoftext|>" , __lowercase=False , **__lowercase , ) -> str: super().__init__( __lowercase , __lowercase , tokenizer_file=__lowercase , unk_token=__lowercase , bos_token=__lowercase , eos_token=__lowercase , add_prefix_space=__lowercase , **__lowercase , ) if kwargs.pop('''add_bos_token''' , __lowercase ): lowerCAmelCase_ : Tuple = kwargs.pop('''name_or_path''' , '''''' ) raise ValueError( '''Currenty GPT2\'s fast tokenizer does NOT support adding a BOS token.''' '''Instead you should use GPT2\'s slow tokenizer class `CodeGenTokenizer` as follows: \n''' f"""`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n""" f"""`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n""" '''This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.''' ''' so that the fast tokenizer works correctly.''' ) lowerCAmelCase_ : Union[str, Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , __lowercase ) != add_prefix_space: lowerCAmelCase_ : Tuple = getattr(__lowercase , pre_tok_state.pop('''type''' ) ) lowerCAmelCase_ : Union[str, Any] = add_prefix_space lowerCAmelCase_ : str = pre_tok_class(**__lowercase ) lowerCAmelCase_ : List[str] = add_prefix_space def lowercase_ ( self , *__lowercase , **__lowercase ) -> BatchEncoding: lowerCAmelCase_ : str = kwargs.get('''is_split_into_words''' , __lowercase ) assert self.add_prefix_space or not is_split_into_words, ( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowercase , **__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> BatchEncoding: lowerCAmelCase_ : Union[str, Any] = kwargs.get('''is_split_into_words''' , __lowercase ) assert self.add_prefix_space or not is_split_into_words, ( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowercase , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: lowerCAmelCase_ : List[str] = self._tokenizer.model.save(__lowercase , name=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False , __lowercase = None , __lowercase = None , **__lowercase , ) -> str: lowerCAmelCase_ : int = super().decode( token_ids=__lowercase , skip_special_tokens=__lowercase , clean_up_tokenization_spaces=__lowercase , **__lowercase , ) if truncate_before_pattern is not None and len(__lowercase ) > 0: lowerCAmelCase_ : Union[str, Any] = self.truncate(__lowercase , __lowercase ) return decoded_text def lowercase_ ( self , __lowercase , __lowercase ) -> Optional[int]: def find_re(__lowercase , __lowercase , __lowercase ): lowerCAmelCase_ : Optional[int] = pattern.search(__lowercase , __lowercase ) return m.start() if m else -1 lowerCAmelCase_ : str = [re.compile(__lowercase , re.MULTILINE ) for pattern in truncate_before_pattern] lowerCAmelCase_ : Any = list(re.finditer('''^print''' , __lowercase , re.MULTILINE ) ) if len(__lowercase ) > 1: lowerCAmelCase_ : Any = completion[: prints[1].start()] lowerCAmelCase_ : Union[str, Any] = list(re.finditer('''^def''' , __lowercase , re.MULTILINE ) ) if len(__lowercase ) > 1: lowerCAmelCase_ : List[Any] = completion[: defs[1].start()] lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Any = [ pos for pos in [find_re(__lowercase , __lowercase , __lowercase ) for terminal in terminals] if pos != -1 ] if len(__lowercase ) > 0: return completion[: min(__lowercase )] else: return completion
262
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = JukeboxTokenizer SCREAMING_SNAKE_CASE__ : int = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowercase_ ( self ) -> Union[str, Any]: import torch lowerCAmelCase_ : Union[str, Any] = JukeboxTokenizer.from_pretrained('''openai/jukebox-1b-lyrics''' ) lowerCAmelCase_ : Any = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase_ ( self ) -> List[Any]: import torch lowerCAmelCase_ : Any = JukeboxTokenizer.from_pretrained('''openai/jukebox-5b-lyrics''' ) lowerCAmelCase_ : str = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : Tuple = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
262
1
def lowerCAmelCase ( lowerCAmelCase_ )-> bool: if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True lowerCAmelCase_ : Optional[Any] = 4 lowerCAmelCase_ : Union[str, Any] = (1 << p) - 1 for _ in range(p - 2 ): lowerCAmelCase_ : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
262
from __future__ import annotations import requests def lowerCAmelCase ( lowerCAmelCase_ )-> dict: lowerCAmelCase_ : List[Any] = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> list[dict]: lowerCAmelCase_ : List[Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCAmelCase_ : Tuple = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> str: lowerCAmelCase_ : Optional[Any] = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join('''* [{title}]({url})'''.format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
262
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _UpperCAmelCase : Dict =logging.get_logger(__name__) _UpperCAmelCase : Any ={"""vocab_file""": """spiece.model"""} _UpperCAmelCase : str ={ """vocab_file""": { """TsinghuaAI/CPM-Generate""": """https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model""", } } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=False , __lowercase=True , __lowercase=False , __lowercase="<s>" , __lowercase="</s>" , __lowercase="<unk>" , __lowercase="<sep>" , __lowercase="<pad>" , __lowercase="<cls>" , __lowercase="<mask>" , __lowercase=["<eop>", "<eod>"] , __lowercase = None , **__lowercase , ) -> None: lowerCAmelCase_ : str = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__lowercase , remove_space=__lowercase , keep_accents=__lowercase , bos_token=__lowercase , eos_token=__lowercase , unk_token=__lowercase , sep_token=__lowercase , pad_token=__lowercase , cls_token=__lowercase , mask_token=__lowercase , additional_special_tokens=__lowercase , sp_model_kwargs=self.sp_model_kwargs , **__lowercase , ) lowerCAmelCase_ : Dict = 3 lowerCAmelCase_ : Optional[Any] = do_lower_case lowerCAmelCase_ : Optional[Any] = remove_space lowerCAmelCase_ : Any = keep_accents lowerCAmelCase_ : int = vocab_file lowerCAmelCase_ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowercase ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( '''You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ''' '''See https://pypi.org/project/jieba/ for installation.''' ) lowerCAmelCase_ : str = jieba lowerCAmelCase_ : int = str.maketrans(''' \n''' , '''\u2582\u2583''' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def lowercase_ ( self ) -> Union[str, Any]: return len(self.sp_model ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = {self.convert_ids_to_tokens(__lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Optional[int]: lowerCAmelCase_ : Tuple = self.__dict__.copy() lowerCAmelCase_ : int = None return state def __setstate__( self , __lowercase ) -> Optional[int]: lowerCAmelCase_ : Any = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowerCAmelCase_ : List[Any] = {} lowerCAmelCase_ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase_ ( self , __lowercase ) -> List[str]: if self.remove_space: lowerCAmelCase_ : Optional[int] = ''' '''.join(inputs.strip().split() ) else: lowerCAmelCase_ : Optional[Any] = inputs lowerCAmelCase_ : Union[str, Any] = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: lowerCAmelCase_ : Optional[Any] = unicodedata.normalize('''NFKD''' , __lowercase ) lowerCAmelCase_ : List[str] = ''''''.join([c for c in outputs if not unicodedata.combining(__lowercase )] ) if self.do_lower_case: lowerCAmelCase_ : Tuple = outputs.lower() return outputs def lowercase_ ( self , __lowercase ) -> List[str]: lowerCAmelCase_ : int = self.preprocess_text(__lowercase ) lowerCAmelCase_ : Any = self.sp_model.encode(__lowercase , out_type=__lowercase ) lowerCAmelCase_ : Tuple = [] for piece in pieces: if len(__lowercase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): lowerCAmelCase_ : Tuple = self.sp_model.EncodeAsPieces(piece[:-1].replace(__lowercase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCAmelCase_ : str = cur_pieces[1:] else: lowerCAmelCase_ : Optional[Any] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__lowercase ) else: new_pieces.append(__lowercase ) return new_pieces def lowercase_ ( self , __lowercase ) -> int: return self.sp_model.PieceToId(__lowercase ) def lowercase_ ( self , __lowercase ) -> str: return self.sp_model.IdToPiece(__lowercase ) def lowercase_ ( self , __lowercase ) -> Dict: lowerCAmelCase_ : List[Any] = ''''''.join(__lowercase ).replace(__lowercase , ''' ''' ).strip() return out_string def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowercase_ ( self , __lowercase , __lowercase = None , __lowercase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase ) if token_ids_a is not None: return ([0] * len(__lowercase )) + [1] + ([0] * len(__lowercase )) + [1, 1] return ([0] * len(__lowercase )) + [1, 1] def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : List[Any] = [self.sep_token_id] lowerCAmelCase_ : str = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowercase ) elif not os.path.isfile(self.vocab_file ): with open(__lowercase , '''wb''' ) as fi: lowerCAmelCase_ : Optional[int] = self.sp_model.serialized_model_proto() fi.write(__lowercase ) return (out_vocab_file,) def lowercase_ ( self , *__lowercase , **__lowercase ) -> Tuple: lowerCAmelCase_ : List[Any] = super()._decode(*__lowercase , **__lowercase ) lowerCAmelCase_ : str = text.replace(''' ''' , '''''' ).replace('''\u2582''' , ''' ''' ).replace('''\u2583''' , '''\n''' ) return text
262
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : List[str] =get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCAmelCase : Optional[int] =25_0004 _UpperCAmelCase : Tuple =25_0020 @require_sentencepiece @require_tokenizers class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = MBartTokenizer SCREAMING_SNAKE_CASE__ : Dict = MBartTokenizerFast SCREAMING_SNAKE_CASE__ : Tuple = True SCREAMING_SNAKE_CASE__ : List[str] = True def lowercase_ ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : str = MBartTokenizer(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[int] = MBartTokenizer(__lowercase , keep_accents=__lowercase ) lowerCAmelCase_ : Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCAmelCase_ : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCAmelCase_ : Dict = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowercase_ ( self ) -> Dict: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase_ : int = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase_ : Optional[int] = self.rust_tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : int = self.tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = tempfile.mkdtemp() lowerCAmelCase_ : Union[str, Any] = tokenizer_r.save_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase_ : str = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Tuple = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : int = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Tuple = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Optional[int] = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Optional[int] = tokenizer_p.save_pretrained(__lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Dict = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) @require_torch @require_sentencepiece @require_tokenizers class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = """facebook/mbart-large-en-ro""" SCREAMING_SNAKE_CASE__ : int = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] SCREAMING_SNAKE_CASE__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] SCREAMING_SNAKE_CASE__ : str = [8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2, EN_CODE] @classmethod def lowercase_ ( cls ) -> Optional[int]: lowerCAmelCase_ : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) lowerCAmelCase_ : Optional[Any] = 1 return cls def lowercase_ ( self ) -> Optional[Any]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 2_5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 2_5_0_0_0_4 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 2_5_0_0_2_0 ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Optional[int] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) def lowercase_ ( self ) -> Any: self.assertIn(__lowercase , self.tokenizer.all_special_ids ) lowerCAmelCase_ : Union[str, Any] = [RO_CODE, 8_8_4, 9_0_1_9, 9_6, 9, 9_1_6, 8_6_7_9_2, 3_6, 1_8_7_4_3, 1_5_5_9_6, 5, 2] lowerCAmelCase_ : Tuple = self.tokenizer.decode(__lowercase , skip_special_tokens=__lowercase ) lowerCAmelCase_ : List[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertNotIn(self.tokenizer.eos_token , __lowercase ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Union[str, Any] = ['''this is gunna be a long sentence ''' * 2_0] assert isinstance(src_text[0] , __lowercase ) lowerCAmelCase_ : str = 1_0 lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , max_length=__lowercase , truncation=__lowercase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __lowercase ) self.assertEqual(len(__lowercase ) , __lowercase ) def lowercase_ ( self ) -> int: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [2_5_0_0_2_6, 2_5_0_0_0_1] ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = tempfile.mkdtemp() lowerCAmelCase_ : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = MBartTokenizer.from_pretrained(__lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowercase ) @require_torch def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Tuple = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowercase , return_tensors='''pt''' ) lowerCAmelCase_ : Tuple = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) lowerCAmelCase_ : int = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 1_4) , batch.input_ids.shape ) self.assertEqual((2, 1_4) , batch.attention_mask.shape ) lowerCAmelCase_ : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[Any] = self.tokenizer(self.src_text , padding=__lowercase , truncation=__lowercase , max_length=3 , return_tensors='''pt''' ) lowerCAmelCase_ : Any = self.tokenizer( text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=1_0 , return_tensors='''pt''' ) lowerCAmelCase_ : int = targets['''input_ids'''] lowerCAmelCase_ : Optional[Any] = shift_tokens_right(__lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(__lowercase ) , { # A, test, EOS, en_XX '''input_ids''': [[6_2, 3_0_3_4, 2, 2_5_0_0_0_4]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 2_5_0_0_0_1, } , )
262
1
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf _UpperCAmelCase : Optional[int] =logging.get_logger(__name__) @dataclass class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = [ """no_inference""", """no_cuda""", """no_tpu""", """no_speed""", """no_memory""", """no_env_print""", """no_multi_process""", ] def __init__( self , **__lowercase ) -> Optional[int]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: lowerCAmelCase_ : Optional[int] = deprecated_arg[3:] lowerCAmelCase_ : str = not kwargs.pop(__lowercase ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no-{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) lowerCAmelCase_ : Union[str, Any] = kwargs.pop('''tpu_name''' , self.tpu_name ) lowerCAmelCase_ : Any = kwargs.pop('''device_idx''' , self.device_idx ) lowerCAmelCase_ : Dict = kwargs.pop('''eager_mode''' , self.eager_mode ) lowerCAmelCase_ : Optional[int] = kwargs.pop('''use_xla''' , self.use_xla ) super().__init__(**__lowercase ) SCREAMING_SNAKE_CASE__ : str = field( default=UpperCAmelCase__, metadata={"""help""": """Name of TPU"""}, ) SCREAMING_SNAKE_CASE__ : int = field( default=0, metadata={"""help""": """CPU / GPU device index. Defaults to 0."""}, ) SCREAMING_SNAKE_CASE__ : bool = field(default=UpperCAmelCase__, metadata={"""help""": """Benchmark models in eager model."""} ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={ """help""": """Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.""" }, ) @cached_property def lowercase_ ( self ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ['''tf'''] ) lowerCAmelCase_ : int = None if self.tpu: try: if self.tpu_name: lowerCAmelCase_ : List[Any] = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: lowerCAmelCase_ : Any = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: lowerCAmelCase_ : str = None return tpu @cached_property def lowercase_ ( self ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ['''tf'''] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) lowerCAmelCase_ : Any = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , '''GPU''' ) lowerCAmelCase_ : Optional[int] = tf.distribute.OneDeviceStrategy(device=f"""/gpu:{self.device_idx}""" ) else: tf.config.set_visible_devices([] , '''GPU''' ) # disable GPU lowerCAmelCase_ : List[Any] = tf.distribute.OneDeviceStrategy(device=f"""/cpu:{self.device_idx}""" ) return strategy @property def lowercase_ ( self ) -> bool: requires_backends(self , ['''tf'''] ) return self._setup_tpu is not None @property def lowercase_ ( self ) -> "tf.distribute.Strategy": requires_backends(self , ['''tf'''] ) return self._setup_strategy @property def lowercase_ ( self ) -> Dict: requires_backends(self , ['''tf'''] ) return tf.config.list_physical_devices('''GPU''' ) @property def lowercase_ ( self ) -> int: requires_backends(self , ['''tf'''] ) if self.cuda: return len(self.gpu_list ) return 0 @property def lowercase_ ( self ) -> bool: return self.n_gpu > 0
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
import qiskit def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> qiskit.result.counts.Counts: lowerCAmelCase_ : Tuple = qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register lowerCAmelCase_ : Optional[int] = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator lowerCAmelCase_ : Dict = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(f"""Total count for various states are: {single_qubit_measure(1, 1)}""")
262
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
262
1
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : int SCREAMING_SNAKE_CASE__ : Node | None = None SCREAMING_SNAKE_CASE__ : Node | None = None def lowerCAmelCase ( )-> Node | None: lowerCAmelCase_ : List[Any] = Node(1 ) lowerCAmelCase_ : Optional[Any] = Node(2 ) lowerCAmelCase_ : int = Node(3 ) lowerCAmelCase_ : Optional[Any] = Node(4 ) lowerCAmelCase_ : Union[str, Any] = Node(5 ) return tree def lowerCAmelCase ( lowerCAmelCase_ )-> list[int]: return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowerCAmelCase ( lowerCAmelCase_ )-> list[int]: return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowerCAmelCase ( lowerCAmelCase_ )-> list[int]: return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowerCAmelCase ( lowerCAmelCase_ )-> int: return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowerCAmelCase ( lowerCAmelCase_ )-> Sequence[Node | None]: lowerCAmelCase_ : list[Any] = [] if root is None: return output lowerCAmelCase_ : Tuple = deque([root] ) while process_queue: lowerCAmelCase_ : int = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Sequence[Node | None]: lowerCAmelCase_ : list[Any] = [] def populate_output(lowerCAmelCase_ , lowerCAmelCase_ ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(lowerCAmelCase_ , lowerCAmelCase_ ) return output def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Sequence[Node | None]: lowerCAmelCase_ : list[Any] = [] def populate_output(lowerCAmelCase_ , lowerCAmelCase_ ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(lowerCAmelCase_ , lowerCAmelCase_ ) return output def lowerCAmelCase ( lowerCAmelCase_ )-> Sequence[Node | None] | list[Any]: if root is None: return [] lowerCAmelCase_ : list[Sequence[Node | None]] = [] lowerCAmelCase_ : str = 0 lowerCAmelCase_ : List[Any] = height(lowerCAmelCase_ ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(lowerCAmelCase_ , lowerCAmelCase_ ) ) lowerCAmelCase_ : int = 1 else: output.append(get_nodes_from_right_to_left(lowerCAmelCase_ , lowerCAmelCase_ ) ) lowerCAmelCase_ : Tuple = 0 return output def lowerCAmelCase ( )-> None: # Main function for testing. lowerCAmelCase_ : Optional[Any] = make_tree() print(f"""In-order Traversal: {inorder(lowerCAmelCase_ )}""" ) print(f"""Pre-order Traversal: {preorder(lowerCAmelCase_ )}""" ) print(f"""Post-order Traversal: {postorder(lowerCAmelCase_ )}""" , '''\n''' ) print(f"""Height of Tree: {height(lowerCAmelCase_ )}""" , '''\n''' ) print('''Complete Level Order Traversal: ''' ) print(level_order(lowerCAmelCase_ ) , '''\n''' ) print('''Level-wise order Traversal: ''' ) for level in range(1 , height(lowerCAmelCase_ ) + 1 ): print(f"""Level {level}:""" , get_nodes_from_left_to_right(lowerCAmelCase_ , level=lowerCAmelCase_ ) ) print('''\nZigZag order Traversal: ''' ) print(zigzag(lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
262
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : str ={ """facebook/vit-mae-base""": """https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json""", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """vit_mae""" def __init__( self , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1e-12 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase=3 , __lowercase=True , __lowercase=1_6 , __lowercase=5_1_2 , __lowercase=8 , __lowercase=2_0_4_8 , __lowercase=0.75 , __lowercase=False , **__lowercase , ) -> str: super().__init__(**__lowercase ) lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : Union[str, Any] = image_size lowerCAmelCase_ : Optional[int] = patch_size lowerCAmelCase_ : Tuple = num_channels lowerCAmelCase_ : List[str] = qkv_bias lowerCAmelCase_ : List[Any] = decoder_num_attention_heads lowerCAmelCase_ : int = decoder_hidden_size lowerCAmelCase_ : Optional[int] = decoder_num_hidden_layers lowerCAmelCase_ : Tuple = decoder_intermediate_size lowerCAmelCase_ : Tuple = mask_ratio lowerCAmelCase_ : Any = norm_pix_loss
262
1
from jiwer import compute_measures import datasets _UpperCAmelCase : Optional[Any] ="""\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ _UpperCAmelCase : Optional[Any] ="""\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. """ _UpperCAmelCase : int =""" Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> wer = datasets.load_metric(\"wer\") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case__( datasets.Metric ): '''simple docstring''' def lowercase_ ( self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def lowercase_ ( self , __lowercase=None , __lowercase=None , __lowercase=False ) -> Union[str, Any]: if concatenate_texts: return compute_measures(__lowercase , __lowercase )["wer"] else: lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Optional[Any] = 0 for prediction, reference in zip(__lowercase , __lowercase ): lowerCAmelCase_ : Tuple = compute_measures(__lowercase , __lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
262
def lowerCAmelCase ( lowerCAmelCase_ = 10**9 )-> int: lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : str = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value lowerCAmelCase_ : Any = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f"""{solution() = }""")
262
1
def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number_of_steps > 0 ), f"""number_of_steps needs to be positive integer, your input {number_of_steps}""" if number_of_steps == 1: return 1 lowerCAmelCase_ , lowerCAmelCase_ : str = 1, 1 for _ in range(number_of_steps - 1 ): lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
262
import inspect import unittest class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: try: import diffusers # noqa: F401 except ImportError: assert False def lowercase_ ( self ) -> List[str]: import diffusers from diffusers.dependency_versions_table import deps lowerCAmelCase_ : Any = inspect.getmembers(__lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowerCAmelCase_ : Optional[int] = '''k-diffusion''' elif backend == "invisible_watermark": lowerCAmelCase_ : Dict = '''invisible-watermark''' assert backend in deps, f"""{backend} is not in the deps table!"""
262
1
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _UpperCAmelCase : List[str] =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ={ """google/umt5-small""": """https://huggingface.co/google/umt5-small/resolve/main/config.json""", # See all umt5 models at https://huggingface.co/models?filter=umt5 } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = """umt5""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] def __init__( self , __lowercase=2_5_0_1_1_2 , __lowercase=5_1_2 , __lowercase=6_4 , __lowercase=1_0_2_4 , __lowercase=8 , __lowercase=None , __lowercase=6 , __lowercase=3_2 , __lowercase=1_2_8 , __lowercase=0.1 , __lowercase=1e-6 , __lowercase=1.0 , __lowercase="gated-gelu" , __lowercase=True , __lowercase=True , __lowercase="T5Tokenizer" , __lowercase=True , __lowercase=0 , __lowercase=1 , __lowercase=0 , **__lowercase , ) -> str: super().__init__( is_encoder_decoder=__lowercase , tokenizer_class=__lowercase , tie_word_embeddings=__lowercase , pad_token_id=__lowercase , eos_token_id=__lowercase , decoder_start_token_id=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_size lowerCAmelCase_ : Optional[Any] = d_model lowerCAmelCase_ : Tuple = d_kv lowerCAmelCase_ : List[str] = d_ff lowerCAmelCase_ : Union[str, Any] = num_layers lowerCAmelCase_ : Dict = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase_ : Dict = num_heads lowerCAmelCase_ : int = relative_attention_num_buckets lowerCAmelCase_ : str = relative_attention_max_distance lowerCAmelCase_ : int = dropout_rate lowerCAmelCase_ : Optional[Any] = layer_norm_epsilon lowerCAmelCase_ : Tuple = initializer_factor lowerCAmelCase_ : List[str] = feed_forward_proj lowerCAmelCase_ : str = use_cache lowerCAmelCase_ : List[str] = self.feed_forward_proj.split('''-''' ) lowerCAmelCase_ : int = act_info[-1] lowerCAmelCase_ : List[Any] = act_info[0] == '''gated''' if len(__lowercase ) > 1 and act_info[0] != "gated" or len(__lowercase ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) if feed_forward_proj == "gated-gelu": lowerCAmelCase_ : int = '''gelu_new''' @property def lowercase_ ( self ) -> List[str]: return self.d_model @property def lowercase_ ( self ) -> Optional[Any]: return self.num_heads @property def lowercase_ ( self ) -> Optional[int]: return self.num_layers class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: lowerCAmelCase_ : List[str] = { '''input_ids''': {0: '''batch''', 1: '''encoder_sequence'''}, '''attention_mask''': {0: '''batch''', 1: '''encoder_sequence'''}, } if self.use_past: lowerCAmelCase_ : List[str] = '''past_encoder_sequence + sequence''' lowerCAmelCase_ : int = {0: '''batch'''} lowerCAmelCase_ : List[str] = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowerCAmelCase_ : str = {0: '''batch''', 1: '''decoder_sequence'''} lowerCAmelCase_ : int = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction='''inputs''' ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def lowercase_ ( self ) -> int: return 1_3 @property def lowercase_ ( self ) -> float: return 5e-4
262
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCAmelCase : Any =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
1
def lowerCAmelCase ( lowerCAmelCase_ = 600_851_475_143 )-> int: try: lowerCAmelCase_ : Optional[int] = int(lowerCAmelCase_ ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Optional[int] = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 lowerCAmelCase_ : Any = i while n % i == 0: lowerCAmelCase_ : Tuple = n // i i += 1 return int(lowerCAmelCase_ ) if __name__ == "__main__": print(f"""{solution() = }""")
262
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
1
_UpperCAmelCase : dict[str, float] ={ "km/h": 1.0, "m/s": 3.6, "mph": 1.60_9344, "knot": 1.852, } _UpperCAmelCase : dict[str, float] ={ "km/h": 1.0, "m/s": 0.2_7777_7778, "mph": 0.6_2137_1192, "knot": 0.5_3995_6803, } def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> float: if unit_to not in speed_chart or unit_from not in speed_chart_inverse: lowerCAmelCase_ : Any = ( f"""Incorrect 'from_type' or 'to_type' value: {unit_from!r}, {unit_to!r}\n""" f"""Valid values are: {", ".join(lowerCAmelCase_ )}""" ) raise ValueError(lowerCAmelCase_ ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
262
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
1
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowercase_ ( self ) -> int: lowerCAmelCase_ : List[str] = 1 lowerCAmelCase_ : List[Any] = 3 lowerCAmelCase_ : Optional[int] = (3_2, 3_2) lowerCAmelCase_ : List[str] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__lowercase ) return image @property def lowercase_ ( self ) -> Optional[int]: torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , ) return model @property def lowercase_ ( self ) -> int: torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def lowercase_ ( self ) -> Any: torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_0_0_6 , ) return RobertaSeriesModelWithTransformation(__lowercase ) @property def lowercase_ ( self ) -> List[Any]: def extract(*__lowercase , **__lowercase ): class snake_case__: '''simple docstring''' def __init__( self ) -> Any: lowerCAmelCase_ : Optional[int] = torch.ones([0] ) def lowercase_ ( self , __lowercase ) -> str: self.pixel_values.to(__lowercase ) return self return Out() return extract def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Union[str, Any] = self.dummy_cond_unet lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=__lowercase ) lowerCAmelCase_ : Union[str, Any] = self.dummy_vae lowerCAmelCase_ : Dict = self.dummy_text_encoder lowerCAmelCase_ : Optional[int] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) lowerCAmelCase_ : Optional[int] = 7_7 lowerCAmelCase_ : List[Any] = self.dummy_image.to(__lowercase ) lowerCAmelCase_ : Union[str, Any] = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk lowerCAmelCase_ : str = AltDiffusionImgaImgPipeline( unet=__lowercase , scheduler=__lowercase , vae=__lowercase , text_encoder=__lowercase , tokenizer=__lowercase , safety_checker=__lowercase , feature_extractor=self.dummy_extractor , ) lowerCAmelCase_ : str = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowercase ) lowerCAmelCase_ : List[str] = alt_pipe.to(__lowercase ) alt_pipe.set_progress_bar_config(disable=__lowercase ) lowerCAmelCase_ : Optional[int] = '''A painting of a squirrel eating a burger''' lowerCAmelCase_ : int = torch.Generator(device=__lowercase ).manual_seed(0 ) lowerCAmelCase_ : str = alt_pipe( [prompt] , generator=__lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=__lowercase , ) lowerCAmelCase_ : Optional[Any] = output.images lowerCAmelCase_ : List[str] = torch.Generator(device=__lowercase ).manual_seed(0 ) lowerCAmelCase_ : Dict = alt_pipe( [prompt] , generator=__lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=__lowercase , return_dict=__lowercase , )[0] lowerCAmelCase_ : str = image[0, -3:, -3:, -1] lowerCAmelCase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowerCAmelCase_ : Any = np.array([0.44_27, 0.37_31, 0.42_49, 0.49_41, 0.45_46, 0.41_48, 0.41_93, 0.46_66, 0.44_99] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Tuple = self.dummy_cond_unet lowerCAmelCase_ : List[str] = PNDMScheduler(skip_prk_steps=__lowercase ) lowerCAmelCase_ : int = self.dummy_vae lowerCAmelCase_ : List[Any] = self.dummy_text_encoder lowerCAmelCase_ : Any = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) lowerCAmelCase_ : List[Any] = 7_7 lowerCAmelCase_ : Tuple = self.dummy_image.to(__lowercase ) # put models in fp16 lowerCAmelCase_ : Union[str, Any] = unet.half() lowerCAmelCase_ : Any = vae.half() lowerCAmelCase_ : int = bert.half() # make sure here that pndm scheduler skips prk lowerCAmelCase_ : str = AltDiffusionImgaImgPipeline( unet=__lowercase , scheduler=__lowercase , vae=__lowercase , text_encoder=__lowercase , tokenizer=__lowercase , safety_checker=__lowercase , feature_extractor=self.dummy_extractor , ) lowerCAmelCase_ : List[Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowercase ) lowerCAmelCase_ : Dict = alt_pipe.to(__lowercase ) alt_pipe.set_progress_bar_config(disable=__lowercase ) lowerCAmelCase_ : Union[str, Any] = '''A painting of a squirrel eating a burger''' lowerCAmelCase_ : Tuple = torch.manual_seed(0 ) lowerCAmelCase_ : Dict = alt_pipe( [prompt] , generator=__lowercase , num_inference_steps=2 , output_type='''np''' , image=__lowercase , ).images assert image.shape == (1, 3_2, 3_2, 3) @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) # resize to resolution that is divisible by 8 but not 16 or 32 lowerCAmelCase_ : Union[str, Any] = init_image.resize((7_6_0, 5_0_4) ) lowerCAmelCase_ : int = '''BAAI/AltDiffusion''' lowerCAmelCase_ : Optional[Any] = AltDiffusionImgaImgPipeline.from_pretrained( __lowercase , safety_checker=__lowercase , ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) pipe.enable_attention_slicing() lowerCAmelCase_ : int = '''A fantasy landscape, trending on artstation''' lowerCAmelCase_ : Tuple = torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = pipe( prompt=__lowercase , image=__lowercase , strength=0.75 , guidance_scale=7.5 , generator=__lowercase , output_type='''np''' , ) lowerCAmelCase_ : Any = output.images[0] lowerCAmelCase_ : Dict = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 7_6_0, 3) lowerCAmelCase_ : Tuple = np.array([0.93_58, 0.93_97, 0.95_99, 0.99_01, 1.00_00, 1.00_00, 0.98_82, 1.00_00, 1.00_00] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> str: lowerCAmelCase_ : Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase_ : List[Any] = init_image.resize((7_6_8, 5_1_2) ) lowerCAmelCase_ : str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''' ) lowerCAmelCase_ : Tuple = '''BAAI/AltDiffusion''' lowerCAmelCase_ : List[Any] = AltDiffusionImgaImgPipeline.from_pretrained( __lowercase , safety_checker=__lowercase , ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) pipe.enable_attention_slicing() lowerCAmelCase_ : Dict = '''A fantasy landscape, trending on artstation''' lowerCAmelCase_ : int = torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = pipe( prompt=__lowercase , image=__lowercase , strength=0.75 , guidance_scale=7.5 , generator=__lowercase , output_type='''np''' , ) lowerCAmelCase_ : Optional[Any] = output.images[0] assert image.shape == (5_1_2, 7_6_8, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
262
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
1
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : List[Any] =logging.get_logger(__name__) _UpperCAmelCase : int ={ """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """bart""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , __lowercase=5_0_2_6_5 , __lowercase=1_0_2_4 , __lowercase=1_2 , __lowercase=4_0_9_6 , __lowercase=1_6 , __lowercase=1_2 , __lowercase=4_0_9_6 , __lowercase=1_6 , __lowercase=0.0 , __lowercase=0.0 , __lowercase="gelu" , __lowercase=1_0_2_4 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=0.0 , __lowercase=False , __lowercase=True , __lowercase=3 , __lowercase=1 , __lowercase=0 , __lowercase=2 , __lowercase=True , __lowercase=2 , __lowercase=2 , **__lowercase , ) -> Dict: lowerCAmelCase_ : List[str] = vocab_size lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : Tuple = d_model lowerCAmelCase_ : Dict = encoder_ffn_dim lowerCAmelCase_ : Optional[int] = encoder_layers lowerCAmelCase_ : Optional[int] = encoder_attention_heads lowerCAmelCase_ : Any = decoder_ffn_dim lowerCAmelCase_ : str = decoder_layers lowerCAmelCase_ : Union[str, Any] = decoder_attention_heads lowerCAmelCase_ : List[str] = dropout lowerCAmelCase_ : Any = attention_dropout lowerCAmelCase_ : int = activation_dropout lowerCAmelCase_ : str = activation_function lowerCAmelCase_ : Dict = init_std lowerCAmelCase_ : Dict = encoder_layerdrop lowerCAmelCase_ : Union[str, Any] = decoder_layerdrop lowerCAmelCase_ : List[str] = classifier_dropout lowerCAmelCase_ : Union[str, Any] = use_cache lowerCAmelCase_ : Union[str, Any] = encoder_layers lowerCAmelCase_ : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__lowercase , pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase , is_encoder_decoder=__lowercase , decoder_start_token_id=__lowercase , forced_eos_token_id=__lowercase , **__lowercase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , __lowercase ): lowerCAmelCase_ : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ '''The config can simply be saved and uploaded again to be fixed.''' ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @property def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : List[str] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowerCAmelCase_ : int = {0: '''batch'''} lowerCAmelCase_ : Optional[int] = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowerCAmelCase_ : Tuple = {0: '''batch''', 1: '''decoder_sequence'''} lowerCAmelCase_ : List[str] = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCAmelCase_ : Tuple = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = self.num_layers for i in range(__lowercase ): lowerCAmelCase_ : Optional[int] = {0: '''batch''', 2: '''past_sequence + sequence'''} lowerCAmelCase_ : Optional[int] = {0: '''batch''', 2: '''past_sequence + sequence'''} else: lowerCAmelCase_ : Optional[int] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : List[str] = super().outputs else: lowerCAmelCase_ : Optional[int] = super(__lowercase , self ).outputs if self.use_past: lowerCAmelCase_ , lowerCAmelCase_ : int = self.num_layers for i in range(__lowercase ): lowerCAmelCase_ : str = {0: '''batch''', 2: '''past_sequence + sequence'''} lowerCAmelCase_ : str = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: lowerCAmelCase_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Generate decoder inputs lowerCAmelCase_ : Any = seq_length if not self.use_past else 1 lowerCAmelCase_ : Dict = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} lowerCAmelCase_ : Dict = dict(**__lowercase , **__lowercase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : str = common_inputs['''input_ids'''].shape lowerCAmelCase_ : Union[str, Any] = common_inputs['''decoder_input_ids'''].shape[1] lowerCAmelCase_ , lowerCAmelCase_ : Tuple = self.num_attention_heads lowerCAmelCase_ : Union[str, Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCAmelCase_ : str = decoder_seq_length + 3 lowerCAmelCase_ : Optional[Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCAmelCase_ : Dict = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(__lowercase , __lowercase )] , dim=1 ) lowerCAmelCase_ : Dict = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.num_layers lowerCAmelCase_ : Tuple = min(__lowercase , __lowercase ) lowerCAmelCase_ : Any = max(__lowercase , __lowercase ) - min_num_layers lowerCAmelCase_ : int = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(__lowercase ): common_inputs["past_key_values"].append( ( torch.zeros(__lowercase ), torch.zeros(__lowercase ), torch.zeros(__lowercase ), torch.zeros(__lowercase ), ) ) # TODO: test this. lowerCAmelCase_ : int = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(__lowercase , __lowercase ): common_inputs["past_key_values"].append((torch.zeros(__lowercase ), torch.zeros(__lowercase )) ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: lowerCAmelCase_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : Dict = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowerCAmelCase_ : int = seqlen + 2 lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self.num_layers lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self.num_attention_heads lowerCAmelCase_ : Any = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCAmelCase_ : str = common_inputs['''attention_mask'''].dtype lowerCAmelCase_ : Tuple = torch.cat( [common_inputs['''attention_mask'''], torch.ones(__lowercase , __lowercase , dtype=__lowercase )] , dim=1 ) lowerCAmelCase_ : int = [ (torch.zeros(__lowercase ), torch.zeros(__lowercase )) for _ in range(__lowercase ) ] return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowerCAmelCase_ : Optional[int] = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCAmelCase_ : Union[str, Any] = tokenizer.num_special_tokens_to_add(__lowercase ) lowerCAmelCase_ : int = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowercase ) # Generate dummy inputs according to compute batch and sequence lowerCAmelCase_ : List[Any] = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCAmelCase_ : Any = dict(tokenizer(__lowercase , return_tensors=__lowercase ) ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : List[str] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase ) elif self.task == "causal-lm": lowerCAmelCase_ : Dict = self._generate_dummy_inputs_for_causal_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase ) else: lowerCAmelCase_ : Tuple = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[Any]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : Union[str, Any] = super()._flatten_past_key_values_(__lowercase , __lowercase , __lowercase , __lowercase ) else: lowerCAmelCase_ : Tuple = super(__lowercase , self )._flatten_past_key_values_( __lowercase , __lowercase , __lowercase , __lowercase )
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
import argparse import copy def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: lowerCAmelCase_ : int = {} with open(lowerCAmelCase_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowerCAmelCase_ : Optional[int] = [] _list.append([line.split()[1], line.split()[2]] ) lowerCAmelCase_ : Union[str, Any] = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowerCAmelCase_ : Optional[int] = [] _list.append([line.split()[0], line.split()[2]] ) lowerCAmelCase_ : List[Any] = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: with open(lowerCAmelCase_ ) as f: lowerCAmelCase_ : Tuple = f.read(1 ) lowerCAmelCase_ : List[str] = start_node lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Tuple = start_node lowerCAmelCase_ : List[str] = 0 while visiting not in first_solution: lowerCAmelCase_ : List[Any] = 10_000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCAmelCase_ ) and k[0] not in first_solution: lowerCAmelCase_ : Dict = k[1] lowerCAmelCase_ : Optional[Any] = k[0] first_solution.append(lowerCAmelCase_ ) lowerCAmelCase_ : int = distance_of_first_solution + int(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = best_node first_solution.append(lowerCAmelCase_ ) lowerCAmelCase_ : str = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowerCAmelCase_ : Optional[int] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10_000 ) return first_solution, distance_of_first_solution def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> int: lowerCAmelCase_ : str = [] for n in solution[1:-1]: lowerCAmelCase_ : str = solution.index(lowerCAmelCase_ ) for kn in solution[1:-1]: lowerCAmelCase_ : List[str] = solution.index(lowerCAmelCase_ ) if n == kn: continue lowerCAmelCase_ : List[str] = copy.deepcopy(lowerCAmelCase_ ) lowerCAmelCase_ : int = kn lowerCAmelCase_ : str = n lowerCAmelCase_ : List[Any] = 0 for k in _tmp[:-1]: lowerCAmelCase_ : Dict = _tmp[_tmp.index(lowerCAmelCase_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowerCAmelCase_ : Union[str, Any] = distance + int(i[1] ) _tmp.append(lowerCAmelCase_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowerCAmelCase_ : str = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowerCAmelCase_ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[Any]: lowerCAmelCase_ : Optional[Any] = 1 lowerCAmelCase_ : List[Any] = first_solution lowerCAmelCase_ : Optional[Any] = [] lowerCAmelCase_ : List[Any] = distance_of_first_solution lowerCAmelCase_ : Tuple = solution while count <= iters: lowerCAmelCase_ : str = find_neighborhood(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Tuple = neighborhood[index_of_best_solution] lowerCAmelCase_ : Any = len(lowerCAmelCase_ ) - 1 lowerCAmelCase_ : Union[str, Any] = False while not found: lowerCAmelCase_ : Union[str, Any] = 0 while i < len(lowerCAmelCase_ ): if best_solution[i] != solution[i]: lowerCAmelCase_ : Dict = best_solution[i] lowerCAmelCase_ : Tuple = solution[i] break lowerCAmelCase_ : List[Any] = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : List[str] = best_solution[:-1] lowerCAmelCase_ : Any = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowerCAmelCase_ : str = cost lowerCAmelCase_ : int = solution else: lowerCAmelCase_ : Dict = index_of_best_solution + 1 lowerCAmelCase_ : Union[str, Any] = neighborhood[index_of_best_solution] if len(lowerCAmelCase_ ) >= size: tabu_list.pop(0 ) lowerCAmelCase_ : Optional[int] = count + 1 return best_solution_ever, best_cost def lowerCAmelCase ( lowerCAmelCase_=None )-> Union[str, Any]: lowerCAmelCase_ : Optional[int] = generate_neighbours(args.File ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = generate_first_solution( args.File , lowerCAmelCase_ ) lowerCAmelCase_ , lowerCAmelCase_ : str = tabu_search( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _UpperCAmelCase : Dict =argparse.ArgumentParser(description="""Tabu Search""") parser.add_argument( """-f""", """--File""", type=str, help="""Path to the file containing the data""", required=True, ) parser.add_argument( """-i""", """--Iterations""", type=int, help="""How many iterations the algorithm should perform""", required=True, ) parser.add_argument( """-s""", """--Size""", type=int, help="""Size of the tabu list""", required=True ) # Pass the arguments to main method main(parser.parse_args())
262
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Tuple =logging.get_logger(__name__) _UpperCAmelCase : str ={ """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = """git_vision_model""" def __init__( self , __lowercase=7_6_8 , __lowercase=3_0_7_2 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase="quick_gelu" , __lowercase=1e-5 , __lowercase=0.0 , __lowercase=0.02 , **__lowercase , ) -> int: super().__init__(**__lowercase ) lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : List[str] = intermediate_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : Optional[Any] = num_attention_heads lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : int = patch_size lowerCAmelCase_ : Any = image_size lowerCAmelCase_ : List[Any] = initializer_range lowerCAmelCase_ : str = attention_dropout lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : Optional[Any] = hidden_act @classmethod def lowercase_ ( cls , __lowercase , **__lowercase ) -> "PretrainedConfig": cls._set_token_in_kwargs(__lowercase ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": lowerCAmelCase_ : Any = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowercase , **__lowercase ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = """git""" def __init__( self , __lowercase=None , __lowercase=3_0_5_2_2 , __lowercase=7_6_8 , __lowercase=6 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1_0_2_4 , __lowercase=0.02 , __lowercase=1e-12 , __lowercase=0 , __lowercase="absolute" , __lowercase=True , __lowercase=False , __lowercase=1_0_1 , __lowercase=1_0_2 , __lowercase=None , **__lowercase , ) -> str: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: lowerCAmelCase_ : Tuple = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) lowerCAmelCase_ : Optional[int] = GitVisionConfig(**__lowercase ) lowerCAmelCase_ : Any = vocab_size lowerCAmelCase_ : Optional[Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : Optional[Any] = intermediate_size lowerCAmelCase_ : Dict = hidden_dropout_prob lowerCAmelCase_ : List[Any] = attention_probs_dropout_prob lowerCAmelCase_ : Optional[Any] = max_position_embeddings lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Optional[int] = layer_norm_eps lowerCAmelCase_ : Optional[Any] = position_embedding_type lowerCAmelCase_ : Union[str, Any] = use_cache lowerCAmelCase_ : Dict = tie_word_embeddings lowerCAmelCase_ : List[str] = num_image_with_embedding lowerCAmelCase_ : List[str] = bos_token_id lowerCAmelCase_ : str = eos_token_id def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : int = copy.deepcopy(self.__dict__ ) lowerCAmelCase_ : Dict = self.vision_config.to_dict() lowerCAmelCase_ : Tuple = self.__class__.model_type return output
262
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
1
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") _UpperCAmelCase : Tuple =logging.getLogger(__name__) @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""}, ) SCREAMING_SNAKE_CASE__ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[str] = field(default=UpperCAmelCase__, metadata={"""help""": """The input training data file (a text file)."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""}, ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={"""help""": """The number of processes to use for the preprocessing."""}, ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) def lowercase_ ( self ) -> List[Any]: if self.train_file is not None: lowerCAmelCase_ : Tuple = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: lowerCAmelCase_ : str = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : PreTrainedTokenizerBase SCREAMING_SNAKE_CASE__ : Union[bool, str, PaddingStrategy] = True SCREAMING_SNAKE_CASE__ : Optional[int] = None SCREAMING_SNAKE_CASE__ : Optional[int] = None def __call__( self , __lowercase ) -> int: lowerCAmelCase_ : Any = '''label''' if '''label''' in features[0].keys() else '''labels''' lowerCAmelCase_ : Tuple = [feature.pop(__lowercase ) for feature in features] lowerCAmelCase_ : str = len(__lowercase ) lowerCAmelCase_ : str = len(features[0]['''input_ids'''] ) lowerCAmelCase_ : Union[str, Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(__lowercase )] for feature in features ] lowerCAmelCase_ : str = list(chain(*__lowercase ) ) lowerCAmelCase_ : Optional[Any] = self.tokenizer.pad( __lowercase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten lowerCAmelCase_ : Any = {k: v.view(__lowercase , __lowercase , -1 ) for k, v in batch.items()} # Add back labels lowerCAmelCase_ : Tuple = torch.tensor(__lowercase , dtype=torch.intaa ) return batch def lowerCAmelCase ( )-> Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Union[str, Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Tuple = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , lowerCAmelCase_ , lowerCAmelCase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCAmelCase_ : Dict = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase_ ) datasets.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCAmelCase_ : Dict = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase_ : str = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: lowerCAmelCase_ : Union[str, Any] = {} if data_args.train_file is not None: lowerCAmelCase_ : int = data_args.train_file if data_args.validation_file is not None: lowerCAmelCase_ : Any = data_args.validation_file lowerCAmelCase_ : Union[str, Any] = data_args.train_file.split('''.''' )[-1] lowerCAmelCase_ : List[str] = load_dataset( lowerCAmelCase_ , data_files=lowerCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. lowerCAmelCase_ : str = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase_ : Optional[int] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCAmelCase_ : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCAmelCase_ : List[Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. lowerCAmelCase_ : Union[str, Any] = [f"""ending{i}""" for i in range(4 )] lowerCAmelCase_ : Union[str, Any] = '''sent1''' lowerCAmelCase_ : Optional[Any] = '''sent2''' if data_args.max_seq_length is None: lowerCAmelCase_ : List[Any] = tokenizer.model_max_length if max_seq_length > 1_024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) lowerCAmelCase_ : Dict = 1_024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) lowerCAmelCase_ : Any = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = [[context] * 4 for context in examples[context_name]] lowerCAmelCase_ : Tuple = examples[question_header_name] lowerCAmelCase_ : Optional[Any] = [ [f"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(lowerCAmelCase_ ) ] # Flatten out lowerCAmelCase_ : List[Any] = list(chain(*lowerCAmelCase_ ) ) lowerCAmelCase_ : Optional[Any] = list(chain(*lowerCAmelCase_ ) ) # Tokenize lowerCAmelCase_ : Any = tokenizer( lowerCAmelCase_ , lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowerCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) lowerCAmelCase_ : Union[str, Any] = raw_datasets['''train'''] if data_args.max_train_samples is not None: lowerCAmelCase_ : Optional[int] = min(len(lowerCAmelCase_ ) , data_args.max_train_samples ) lowerCAmelCase_ : List[Any] = train_dataset.select(range(lowerCAmelCase_ ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): lowerCAmelCase_ : int = train_dataset.map( lowerCAmelCase_ , batched=lowerCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) lowerCAmelCase_ : Tuple = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: lowerCAmelCase_ : List[str] = min(len(lowerCAmelCase_ ) , data_args.max_eval_samples ) lowerCAmelCase_ : List[str] = eval_dataset.select(range(lowerCAmelCase_ ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): lowerCAmelCase_ : List[Any] = eval_dataset.map( lowerCAmelCase_ , batched=lowerCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator lowerCAmelCase_ : Dict = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowerCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowerCAmelCase_ ): lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = eval_predictions lowerCAmelCase_ : int = np.argmax(lowerCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer lowerCAmelCase_ : int = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , compute_metrics=lowerCAmelCase_ , ) # Training if training_args.do_train: lowerCAmelCase_ : Any = None if training_args.resume_from_checkpoint is not None: lowerCAmelCase_ : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCAmelCase_ : Any = last_checkpoint lowerCAmelCase_ : Optional[int] = trainer.train(resume_from_checkpoint=lowerCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCAmelCase_ : List[str] = train_result.metrics lowerCAmelCase_ : Optional[Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCAmelCase_ ) ) lowerCAmelCase_ : str = min(lowerCAmelCase_ , len(lowerCAmelCase_ ) ) trainer.log_metrics('''train''' , lowerCAmelCase_ ) trainer.save_metrics('''train''' , lowerCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCAmelCase_ : List[str] = trainer.evaluate() lowerCAmelCase_ : int = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = min(lowerCAmelCase_ , len(lowerCAmelCase_ ) ) trainer.log_metrics('''eval''' , lowerCAmelCase_ ) trainer.save_metrics('''eval''' , lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase_ ) else: trainer.create_model_card(**lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> int: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
1
import inspect import unittest class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: try: import diffusers # noqa: F401 except ImportError: assert False def lowercase_ ( self ) -> List[str]: import diffusers from diffusers.dependency_versions_table import deps lowerCAmelCase_ : Any = inspect.getmembers(__lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowerCAmelCase_ : Optional[int] = '''k-diffusion''' elif backend == "invisible_watermark": lowerCAmelCase_ : Dict = '''invisible-watermark''' assert backend in deps, f"""{backend} is not in the deps table!"""
262
from __future__ import annotations import math class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : str = size # approximate the overall size of segment tree with given value lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] lowerCAmelCase_ : Optional[int] = [0 for i in range(0 , 4 * size )] # flag for lazy update def lowercase_ ( self , __lowercase ) -> int: return idx * 2 def lowercase_ ( self , __lowercase ) -> int: return idx * 2 + 1 def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> None: if left_element == right_element: lowerCAmelCase_ : Tuple = a[left_element - 1] else: lowerCAmelCase_ : int = (left_element + right_element) // 2 self.build(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase ) self.build(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase ) lowerCAmelCase_ : Any = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> bool: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Union[str, Any] = False if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Any = self.lazy[idx] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCAmelCase_ : Dict = val if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = val lowerCAmelCase_ : List[Any] = val lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : List[str] = True return True lowerCAmelCase_ : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) self.update(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) return True def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> int | float: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Optional[Any] = False if left_element != right_element: lowerCAmelCase_ : List[Any] = self.lazy[idx] lowerCAmelCase_ : Dict = self.lazy[idx] lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCAmelCase_ : List[Any] = (left_element + right_element) // 2 lowerCAmelCase_ : Tuple = self.query(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.query(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase ) return max(__lowercase , __lowercase ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , __lowercase , __lowercase ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _UpperCAmelCase : str =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _UpperCAmelCase : List[str] =15 _UpperCAmelCase : Any =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
262
1
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class snake_case__: '''simple docstring''' def __init__( self , __lowercase , __lowercase=3 , __lowercase=3_2 , __lowercase=3 , __lowercase=1_0 , __lowercase=[8, 1_6, 3_2, 6_4] , __lowercase=[1, 1, 2, 1] , __lowercase=True , __lowercase=True , __lowercase="relu" , __lowercase=3 , __lowercase=None , __lowercase=["stage2", "stage3", "stage4"] , __lowercase=[2, 3, 4] , __lowercase=1 , ) -> List[Any]: lowerCAmelCase_ : Dict = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : List[str] = image_size lowerCAmelCase_ : Any = num_channels lowerCAmelCase_ : List[str] = embeddings_size lowerCAmelCase_ : List[str] = hidden_sizes lowerCAmelCase_ : Optional[int] = depths lowerCAmelCase_ : str = is_training lowerCAmelCase_ : Union[str, Any] = use_labels lowerCAmelCase_ : Optional[Any] = hidden_act lowerCAmelCase_ : Optional[Any] = num_labels lowerCAmelCase_ : Union[str, Any] = scope lowerCAmelCase_ : List[Any] = len(__lowercase ) lowerCAmelCase_ : Union[str, Any] = out_features lowerCAmelCase_ : List[Any] = out_indices lowerCAmelCase_ : int = num_groups def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase_ : List[str] = None if self.use_labels: lowerCAmelCase_ : Any = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase_ : List[Any] = self.get_config() return config, pixel_values, labels def lowercase_ ( self ) -> str: return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Any: lowerCAmelCase_ : Dict = BitModel(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : List[str] = model(__lowercase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.num_labels lowerCAmelCase_ : int = BitForImageClassification(__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Tuple = model(__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Tuple = BitBackbone(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : List[str] = model(__lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowerCAmelCase_ : Any = None lowerCAmelCase_ : Optional[Any] = BitBackbone(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Any = model(__lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = config_and_inputs lowerCAmelCase_ : Any = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class snake_case__( UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : int = ( {"""feature-extraction""": BitModel, """image-classification""": BitForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : int = False SCREAMING_SNAKE_CASE__ : int = False SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : Any = False SCREAMING_SNAKE_CASE__ : List[Any] = False def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Optional[int] = BitModelTester(self ) lowerCAmelCase_ : str = ConfigTester(self , config_class=__lowercase , has_text_modality=__lowercase ) def lowercase_ ( self ) -> List[str]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase_ ( self ) -> Tuple: return @unittest.skip(reason='''Bit does not output attentions''' ) def lowercase_ ( self ) -> List[Any]: pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def lowercase_ ( self ) -> Optional[int]: pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def lowercase_ ( self ) -> Optional[Any]: pass def lowercase_ ( self ) -> int: lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : Dict = model_class(__lowercase ) lowerCAmelCase_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase_ : Dict = [*signature.parameters.keys()] lowerCAmelCase_ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowercase ) def lowercase_ ( self ) -> str: lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__lowercase ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : str = model_class(config=__lowercase ) for name, module in model.named_modules(): if isinstance(__lowercase , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def lowercase_ ( self ) -> List[Any]: def check_hidden_states_output(__lowercase , __lowercase , __lowercase ): lowerCAmelCase_ : Any = model_class(__lowercase ) model.to(__lowercase ) model.eval() with torch.no_grad(): lowerCAmelCase_ : Tuple = model(**self._prepare_for_class(__lowercase , __lowercase ) ) lowerCAmelCase_ : int = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase_ : Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(__lowercase ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : Any = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowerCAmelCase_ : Any = layer_type lowerCAmelCase_ : List[str] = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase_ : List[Any] = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def lowercase_ ( self ) -> Optional[Any]: pass def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) @slow def lowercase_ ( self ) -> List[str]: for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Optional[int] = BitModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def lowerCAmelCase ( )-> Dict: lowerCAmelCase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class snake_case__( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase_ ( self ) -> Optional[int]: return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowercase_ ( self ) -> str: lowerCAmelCase_ : Optional[Any] = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(__lowercase ) lowerCAmelCase_ : Any = self.default_image_processor lowerCAmelCase_ : int = prepare_img() lowerCAmelCase_ : List[str] = image_processor(images=__lowercase , return_tensors='''pt''' ).to(__lowercase ) # forward pass with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model(**__lowercase ) # verify the logits lowerCAmelCase_ : List[Any] = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowercase ) lowerCAmelCase_ : List[str] = torch.tensor([[-0.65_26, -0.52_63, -1.43_98]] ).to(__lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowercase , atol=1e-4 ) ) @require_torch class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = (BitBackbone,) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Dict = BitConfig SCREAMING_SNAKE_CASE__ : Dict = False def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[str] = BitModelTester(self )
262
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py _UpperCAmelCase : Optional[int] ="""src/transformers""" _UpperCAmelCase : str ="""docs/source/en""" _UpperCAmelCase : Optional[int] =""".""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : int = f.readlines() # Find the start prompt. lowerCAmelCase_ : List[Any] = 0 while not lines[start_index].startswith(lowerCAmelCase_ ): start_index += 1 start_index += 1 lowerCAmelCase_ : List[str] = start_index while not lines[end_index].startswith(lowerCAmelCase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | _UpperCAmelCase : Optional[Any] ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. _UpperCAmelCase : Optional[int] =re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _UpperCAmelCase : Dict =re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _UpperCAmelCase : Optional[Any] =re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : Optional[int] =direct_transformers_import(TRANSFORMERS_PATH) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : str = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' , lowerCAmelCase_ ) return [m.group(0 ) for m in matches] def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Tuple = 2 if text == '''✅''' or text == '''❌''' else len(lowerCAmelCase_ ) lowerCAmelCase_ : int = (width - text_length) // 2 lowerCAmelCase_ : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def lowerCAmelCase ( )-> str: lowerCAmelCase_ : Any = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowerCAmelCase_ : Dict = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } lowerCAmelCase_ : List[Any] = {name: config.replace('''Config''' , '''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. lowerCAmelCase_ : Tuple = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = None if attr_name.endswith('''Tokenizer''' ): lowerCAmelCase_ : Union[str, Any] = slow_tokenizers lowerCAmelCase_ : List[str] = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): lowerCAmelCase_ : int = fast_tokenizers lowerCAmelCase_ : Union[str, Any] = attr_name[:-13] elif _re_tf_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = tf_models lowerCAmelCase_ : str = _re_tf_models.match(lowerCAmelCase_ ).groups()[0] elif _re_flax_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = flax_models lowerCAmelCase_ : Union[str, Any] = _re_flax_models.match(lowerCAmelCase_ ).groups()[0] elif _re_pt_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Any = pt_models lowerCAmelCase_ : List[Any] = _re_pt_models.match(lowerCAmelCase_ ).groups()[0] if lookup_dict is not None: while len(lowerCAmelCase_ ) > 0: if attr_name in model_name_to_prefix.values(): lowerCAmelCase_ : Union[str, Any] = True break # Try again after removing the last word in the name lowerCAmelCase_ : Any = ''''''.join(camel_case_split(lowerCAmelCase_ )[:-1] ) # Let's build that table! lowerCAmelCase_ : int = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) lowerCAmelCase_ : Tuple = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). lowerCAmelCase_ : Union[str, Any] = [len(lowerCAmelCase_ ) + 2 for c in columns] lowerCAmelCase_ : Optional[Any] = max([len(lowerCAmelCase_ ) for name in model_names] ) + 2 # Build the table per se lowerCAmelCase_ : Dict = '''|''' + '''|'''.join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for c, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" lowerCAmelCase_ : List[str] = {True: '''✅''', False: '''❌'''} for name in model_names: lowerCAmelCase_ : List[Any] = model_name_to_prefix[name] lowerCAmelCase_ : Union[str, Any] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for l, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + "|\n" return table def lowerCAmelCase ( lowerCAmelCase_=False )-> Tuple: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = _find_text_in_file( filename=os.path.join(lowerCAmelCase_ , '''index.md''' ) , start_prompt='''<!--This table is updated automatically from the auto modules''' , end_prompt='''<!-- End table-->''' , ) lowerCAmelCase_ : Tuple = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCAmelCase_ , '''index.md''' ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Tuple =parser.parse_args() check_model_table(args.fix_and_overwrite)
262
1
from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig _UpperCAmelCase : Dict =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ="""T5Config""" class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """mt5""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = MTaConfig class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """mt5""" SCREAMING_SNAKE_CASE__ : Optional[int] = MTaConfig class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = """mt5""" SCREAMING_SNAKE_CASE__ : Dict = MTaConfig
262
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer _UpperCAmelCase : Any =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : str ={ """vocab_file""": {"""mobilebert-uncased""": """https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt"""}, """tokenizer_file""": { """mobilebert-uncased""": """https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json""" }, } _UpperCAmelCase : Optional[Any] ={"""mobilebert-uncased""": 512} _UpperCAmelCase : Dict ={} class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Union[str, Any] = MobileBertTokenizer def __init__( self , __lowercase=None , __lowercase=None , __lowercase=True , __lowercase="[UNK]" , __lowercase="[SEP]" , __lowercase="[PAD]" , __lowercase="[CLS]" , __lowercase="[MASK]" , __lowercase=True , __lowercase=None , **__lowercase , ) -> Union[str, Any]: super().__init__( __lowercase , tokenizer_file=__lowercase , do_lower_case=__lowercase , unk_token=__lowercase , sep_token=__lowercase , pad_token=__lowercase , cls_token=__lowercase , mask_token=__lowercase , tokenize_chinese_chars=__lowercase , strip_accents=__lowercase , **__lowercase , ) lowerCAmelCase_ : Optional[int] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , __lowercase ) != do_lower_case or normalizer_state.get('''strip_accents''' , __lowercase ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , __lowercase ) != tokenize_chinese_chars ): lowerCAmelCase_ : Optional[int] = getattr(__lowercase , normalizer_state.pop('''type''' ) ) lowerCAmelCase_ : Optional[Any] = do_lower_case lowerCAmelCase_ : Optional[int] = strip_accents lowerCAmelCase_ : Optional[Any] = tokenize_chinese_chars lowerCAmelCase_ : Union[str, Any] = normalizer_class(**__lowercase ) lowerCAmelCase_ : List[str] = do_lower_case def lowercase_ ( self , __lowercase , __lowercase=None ) -> Any: lowerCAmelCase_ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Tuple = [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: lowerCAmelCase_ : int = self._tokenizer.model.save(__lowercase , name=__lowercase ) return tuple(__lowercase )
262
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : Tuple =None _UpperCAmelCase : int =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Any ={ """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """facebook/nllb-large-en-ro""": 1024, """facebook/nllb-200-distilled-600M""": 1024, } # fmt: off _UpperCAmelCase : Any =["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : int = NllbTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=False , **__lowercase , ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : int = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : List[Any] = legacy_behaviour super().__init__( vocab_file=__lowercase , tokenizer_file=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , additional_special_tokens=__lowercase , legacy_behaviour=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_file lowerCAmelCase_ : List[Any] = False if not self.vocab_file else True lowerCAmelCase_ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) lowerCAmelCase_ : Optional[Any] = { lang_code: self.convert_tokens_to_ids(__lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase_ : Any = src_lang if src_lang is not None else '''eng_Latn''' lowerCAmelCase_ : str = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase_ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowercase_ ( self ) -> str: return self._src_lang @src_lang.setter def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : int = self(__lowercase , add_special_tokens=__lowercase , return_tensors=__lowercase , **__lowercase ) lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) lowerCAmelCase_ : List[Any] = tgt_lang_id return inputs def lowercase_ ( self , __lowercase , __lowercase = "eng_Latn" , __lowercase = None , __lowercase = "fra_Latn" , **__lowercase , ) -> BatchEncoding: lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : List[str] = tgt_lang return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def lowercase_ ( self ) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[str] = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[str] = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Optional[int] = [self.cur_lang_code] lowerCAmelCase_ : List[Any] = [self.eos_token_id] lowerCAmelCase_ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Any = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Any = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Any = [self.cur_lang_code] lowerCAmelCase_ : Any = [self.eos_token_id] lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ): copyfile(self.vocab_file , __lowercase ) return (out_vocab_file,)
262
1
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = BioGptTokenizer SCREAMING_SNAKE_CASE__ : int = False def lowercase_ ( self ) -> Dict: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase_ : List[str] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] lowerCAmelCase_ : Dict = dict(zip(__lowercase , range(len(__lowercase ) ) ) ) lowerCAmelCase_ : List[Any] = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] lowerCAmelCase_ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(__lowercase ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(__lowercase ) ) def lowercase_ ( self , __lowercase ) -> Optional[int]: lowerCAmelCase_ : List[str] = '''lower newer''' lowerCAmelCase_ : str = '''lower newer''' return input_text, output_text def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Tuple = BioGptTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase_ : str = '''lower''' lowerCAmelCase_ : Optional[Any] = ['''low''', '''er</w>'''] lowerCAmelCase_ : Dict = tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) lowerCAmelCase_ : Union[str, Any] = tokens + ['''<unk>'''] lowerCAmelCase_ : Dict = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ) , __lowercase ) @slow def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : str = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) lowerCAmelCase_ : int = tokenizer.encode('''sequence builders''' , add_special_tokens=__lowercase ) lowerCAmelCase_ : str = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__lowercase ) lowerCAmelCase_ : str = tokenizer.build_inputs_with_special_tokens(__lowercase ) lowerCAmelCase_ : Optional[Any] = tokenizer.build_inputs_with_special_tokens(__lowercase , __lowercase ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
262
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _UpperCAmelCase : Optional[Any] =NewType("""DataClass""", Any) _UpperCAmelCase : Dict =NewType("""DataClassType""", Any) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase ( lowerCAmelCase_ )-> Callable[[str], Any]: lowerCAmelCase_ : str = {str(lowerCAmelCase_ ): choice for choice in choices} return lambda lowerCAmelCase_ : str_to_choice.get(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( *, lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = None , **lowerCAmelCase_ , )-> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase_ : Dict = {} if aliases is not None: lowerCAmelCase_ : str = aliases if help is not None: lowerCAmelCase_ : Tuple = help return dataclasses.field(metadata=lowerCAmelCase_ , default=lowerCAmelCase_ , default_factory=lowerCAmelCase_ , **lowerCAmelCase_ ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Iterable[DataClassType] def __init__( self , __lowercase , **__lowercase ) -> List[str]: # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase_ : Optional[int] = ArgumentDefaultsHelpFormatter super().__init__(**__lowercase ) if dataclasses.is_dataclass(__lowercase ): lowerCAmelCase_ : Union[str, Any] = [dataclass_types] lowerCAmelCase_ : List[Any] = list(__lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowercase ) @staticmethod def lowercase_ ( __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[Any] = f"""--{field.name}""" lowerCAmelCase_ : Tuple = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowercase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase_ : List[str] = kwargs.pop('''aliases''' , [] ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[Any] = [aliases] lowerCAmelCase_ : Any = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(__lowercase , '''UnionType''' ) and isinstance(__lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowercase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(__lowercase ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase_ : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase_ : Dict = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase_ : str = ( field.type.__args__[0] if isinstance(__lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase_ : List[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase_ : Dict = {} if origin_type is Literal or (isinstance(field.type , __lowercase ) and issubclass(field.type , __lowercase )): if origin_type is Literal: lowerCAmelCase_ : Optional[Any] = field.type.__args__ else: lowerCAmelCase_ : int = [x.value for x in field.type] lowerCAmelCase_ : str = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase_ : str = field.default else: lowerCAmelCase_ : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase_ : Tuple = copy(__lowercase ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase_ : Dict = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase_ : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase_ : int = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase_ : List[Any] = True elif isclass(__lowercase ) and issubclass(__lowercase , __lowercase ): lowerCAmelCase_ : Union[str, Any] = field.type.__args__[0] lowerCAmelCase_ : Dict = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : Any = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase_ : Dict = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : List[Any] = field.default_factory() else: lowerCAmelCase_ : int = True parser.add_argument(__lowercase , *__lowercase , **__lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase_ : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: if hasattr(__lowercase , '''_argument_group_name''' ): lowerCAmelCase_ : str = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase_ : Dict = self try: lowerCAmelCase_ : Dict[str, type] = get_type_hints(__lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(__lowercase ): lowerCAmelCase_ : Any = '''.'''.join(map(__lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(__lowercase ): if not field.init: continue lowerCAmelCase_ : Optional[int] = type_hints[field.name] self._parse_dataclass_field(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase=None , __lowercase=False , __lowercase=True , __lowercase=None , __lowercase=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase_ : str = [] if args_filename: args_files.append(Path(__lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase_ : str = ArgumentParser() args_file_parser.add_argument(__lowercase , type=__lowercase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = args_file_parser.parse_known_args(args=__lowercase ) lowerCAmelCase_ : int = vars(__lowercase ).get(args_file_flag.lstrip('''-''' ) , __lowercase ) if cmd_args_file_paths: args_files.extend([Path(__lowercase ) for p in cmd_args_file_paths] ) lowerCAmelCase_ : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.parse_known_args(args=__lowercase ) lowerCAmelCase_ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase_ : str = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : str = {k: v for k, v in vars(__lowercase ).items() if k in keys} for k in keys: delattr(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = dtype(**__lowercase ) outputs.append(__lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : int = set(args.keys() ) lowerCAmelCase_ : str = [] for dtype in self.dataclass_types: lowerCAmelCase_ : int = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : List[str] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase_ : List[str] = dtype(**__lowercase ) outputs.append(__lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(__lowercase )}""" ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: with open(Path(__lowercase ) , encoding='''utf-8''' ) as open_json_file: lowerCAmelCase_ : Dict = json.loads(open_json_file.read() ) lowerCAmelCase_ : str = self.parse_dict(__lowercase , allow_extra_keys=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : Optional[Any] = self.parse_dict(yaml.safe_load(Path(__lowercase ).read_text() ) , allow_extra_keys=__lowercase ) return tuple(__lowercase )
262
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""image_processor""", """tokenizer"""] SCREAMING_SNAKE_CASE__ : List[Any] = """ViTImageProcessor""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , __lowercase=None , __lowercase=None , **__lowercase ) -> int: lowerCAmelCase_ : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __lowercase , ) lowerCAmelCase_ : int = kwargs.pop('''feature_extractor''' ) lowerCAmelCase_ : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(__lowercase , __lowercase ) def __call__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , **__lowercase ) -> List[Any]: if text is None and visual_prompt is None and images is None: raise ValueError('''You have to specify either text, visual prompt or images.''' ) if text is not None and visual_prompt is not None: raise ValueError('''You have to specify exactly one type of prompt. Either text or visual prompt.''' ) if text is not None: lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , return_tensors=__lowercase , **__lowercase ) if visual_prompt is not None: lowerCAmelCase_ : Any = self.image_processor(__lowercase , return_tensors=__lowercase , **__lowercase ) if images is not None: lowerCAmelCase_ : Union[str, Any] = self.image_processor(__lowercase , return_tensors=__lowercase , **__lowercase ) if visual_prompt is not None and images is not None: lowerCAmelCase_ : str = { '''pixel_values''': image_features.pixel_values, '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: lowerCAmelCase_ : Dict = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: lowerCAmelCase_ : int = { '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**__lowercase ) , tensor_type=__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> List[str]: return self.tokenizer.batch_decode(*__lowercase , **__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> int: return self.tokenizer.decode(*__lowercase , **__lowercase ) @property def lowercase_ ( self ) -> List[Any]: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __lowercase , ) return self.image_processor_class @property def lowercase_ ( self ) -> List[Any]: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __lowercase , ) return self.image_processor
262
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: return EnvironmentCommand() class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @staticmethod def lowercase_ ( __lowercase ) -> List[Any]: lowerCAmelCase_ : List[str] = parser.add_parser('''env''' ) download_parser.set_defaults(func=__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = huggingface_hub.__version__ lowerCAmelCase_ : str = '''not installed''' lowerCAmelCase_ : str = '''NA''' if is_torch_available(): import torch lowerCAmelCase_ : Any = torch.__version__ lowerCAmelCase_ : str = torch.cuda.is_available() lowerCAmelCase_ : List[str] = '''not installed''' if is_transformers_available(): import transformers lowerCAmelCase_ : Any = transformers.__version__ lowerCAmelCase_ : Optional[Any] = '''not installed''' if is_accelerate_available(): import accelerate lowerCAmelCase_ : List[Any] = accelerate.__version__ lowerCAmelCase_ : List[str] = '''not installed''' if is_xformers_available(): import xformers lowerCAmelCase_ : Optional[Any] = xformers.__version__ lowerCAmelCase_ : int = { '''`diffusers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Huggingface_hub version''': hub_version, '''Transformers version''': transformers_version, '''Accelerate version''': accelerate_version, '''xFormers version''': xformers_version, '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def lowercase_ ( __lowercase ) -> str: return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
262
1
import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( SeqaSeqDataset, calculate_bleu, calculate_rouge, chunks, lmap, load_json, parse_numeric_n_bool_cl_kwargs, save_json, use_task_specific_params, write_txt_file, ) _UpperCAmelCase : int =getLogger(__name__) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 8 , lowerCAmelCase_ = 1_024 , lowerCAmelCase_="val" , lowerCAmelCase_=None , lowerCAmelCase_=False , lowerCAmelCase_="summarization" , lowerCAmelCase_=None , lowerCAmelCase_=1 , lowerCAmelCase_ = None , lowerCAmelCase_="" , **lowerCAmelCase_ , )-> Dict: lowerCAmelCase_ : Optional[Any] = str(lowerCAmelCase_ ) assert local_rank is not None torch.distributed.init_process_group(backend='''nccl''' , rank=lowerCAmelCase_ ) lowerCAmelCase_ : Dict = Path(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = save_dir.joinpath(f"""rank_{local_rank}_output.json""" ) torch.cuda.set_device(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = AutoModelForSeqaSeqLM.from_pretrained(lowerCAmelCase_ ).cuda() if fpaa: lowerCAmelCase_ : Any = model.half() # determine if we need to increase num_beams use_task_specific_params(lowerCAmelCase_ , lowerCAmelCase_ ) # update config with task specific params lowerCAmelCase_ : Optional[Any] = generate_kwargs.pop('''num_beams''' , model.config.num_beams ) # AttributeError risk? if num_return_sequences > num_beams: lowerCAmelCase_ : int = num_return_sequences lowerCAmelCase_ : Optional[Any] = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) logger.info(f"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. if max_source_length is None: lowerCAmelCase_ : int = tokenizer.model_max_length if prefix is None: lowerCAmelCase_ : Tuple = prefix or getattr(model.config , '''prefix''' , '''''' ) or '''''' lowerCAmelCase_ : Optional[int] = SeqaSeqDataset( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , max_target_length=1_024 , type_path=lowerCAmelCase_ , n_obs=lowerCAmelCase_ , prefix=lowerCAmelCase_ , **lowerCAmelCase_ , ) # I set shuffle=True for a more accurate progress bar. # If all the longest samples are first, the prog bar estimate is too high at the beginning. lowerCAmelCase_ : Optional[int] = ds.make_sortish_sampler(lowerCAmelCase_ , distributed=lowerCAmelCase_ , add_extra_examples=lowerCAmelCase_ , shuffle=lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = DataLoader(lowerCAmelCase_ , sampler=lowerCAmelCase_ , batch_size=lowerCAmelCase_ , collate_fn=ds.collate_fn ) lowerCAmelCase_ : Dict = [] for batch in tqdm(lowerCAmelCase_ ): lowerCAmelCase_ : Union[str, Any] = model.generate( input_ids=batch['''input_ids'''].to(model.device ) , attention_mask=batch['''attention_mask'''].to(model.device ) , num_return_sequences=lowerCAmelCase_ , num_beams=lowerCAmelCase_ , **lowerCAmelCase_ , ) lowerCAmelCase_ : Any = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = batch['''ids'''] if num_return_sequences > 1: lowerCAmelCase_ : Optional[int] = chunks(lowerCAmelCase_ , lowerCAmelCase_ ) # batch size chunks, each of size num_return_seq for i, pred in enumerate(lowerCAmelCase_ ): results.append({'''pred''': pred, '''id''': ids[i].item()} ) save_json(lowerCAmelCase_ , lowerCAmelCase_ ) return results, sampler.num_replicas def lowerCAmelCase ( )-> str: lowerCAmelCase_ : int = argparse.ArgumentParser( epilog='''Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate''' ) parser.add_argument('''--data_dir''' , type=lowerCAmelCase_ , help='''like cnn_dm/test.source''' ) parser.add_argument( '''--model_name''' , type=lowerCAmelCase_ , help='''like facebook/bart-large-cnn,t5-base, etc.''' , default='''sshleifer/distilbart-xsum-12-3''' , ) parser.add_argument('''--save_dir''' , type=lowerCAmelCase_ , help='''where to save''' , default='''tmp_gen''' ) parser.add_argument('''--max_source_length''' , type=lowerCAmelCase_ , default=lowerCAmelCase_ ) parser.add_argument( '''--type_path''' , type=lowerCAmelCase_ , default='''test''' , help='''which subset to evaluate typically train/val/test''' ) parser.add_argument('''--task''' , type=lowerCAmelCase_ , default='''summarization''' , help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' , type=lowerCAmelCase_ , default=8 , required=lowerCAmelCase_ , help='''batch size''' ) parser.add_argument( '''--local_rank''' , type=lowerCAmelCase_ , default=-1 , required=lowerCAmelCase_ , help='''should be passed by distributed.launch''' ) parser.add_argument( '''--n_obs''' , type=lowerCAmelCase_ , default=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''How many observations. Defaults to all.''' ) parser.add_argument( '''--num_return_sequences''' , type=lowerCAmelCase_ , default=1 , required=lowerCAmelCase_ , help='''How many sequences to return''' ) parser.add_argument( '''--sync_timeout''' , type=lowerCAmelCase_ , default=600 , required=lowerCAmelCase_ , help='''How long should master process wait for other processes to finish.''' , ) parser.add_argument('''--src_lang''' , type=lowerCAmelCase_ , default=lowerCAmelCase_ , required=lowerCAmelCase_ ) parser.add_argument('''--tgt_lang''' , type=lowerCAmelCase_ , default=lowerCAmelCase_ , required=lowerCAmelCase_ ) parser.add_argument( '''--prefix''' , type=lowerCAmelCase_ , required=lowerCAmelCase_ , default=lowerCAmelCase_ , help='''will be added to the begininng of src examples''' ) parser.add_argument('''--fp16''' , action='''store_true''' ) parser.add_argument('''--debug''' , action='''store_true''' ) lowerCAmelCase_ : Tuple = time.time() lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = parser.parse_known_args() lowerCAmelCase_ : Any = parse_numeric_n_bool_cl_kwargs(lowerCAmelCase_ ) if generate_kwargs and args.local_rank <= 0: print(f"""parsed the following generate kwargs: {generate_kwargs}""" ) lowerCAmelCase_ : int = Path(args.save_dir + '''_tmp''' ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) # this handles locking. lowerCAmelCase_ : Dict = list(json_save_dir.glob('''rank_*.json''' ) ) if intermediate_files: raise ValueError(f"""Found files at {json_save_dir} please move or remove them.""" ) # In theory, a node could finish and save before another node hits this. If this happens, we can address later. lowerCAmelCase_ : Optional[Any] = {} if args.src_lang is not None: lowerCAmelCase_ : Optional[Any] = args.src_lang if args.tgt_lang is not None: lowerCAmelCase_ : Optional[int] = args.tgt_lang Path(args.save_dir ).mkdir(exist_ok=lowerCAmelCase_ ) lowerCAmelCase_ , lowerCAmelCase_ : str = eval_data_dir( args.data_dir , lowerCAmelCase_ , args.model_name , type_path=args.type_path , bs=args.bs , fpaa=args.fpaa , task=args.task , local_rank=args.local_rank , n_obs=args.n_obs , max_source_length=args.max_source_length , num_return_sequences=args.num_return_sequences , prefix=args.prefix , dataset_kwargs=lowerCAmelCase_ , **lowerCAmelCase_ , ) if args.local_rank <= 0: lowerCAmelCase_ : List[str] = Path(args.save_dir ) save_dir.mkdir(exist_ok=lowerCAmelCase_ ) lowerCAmelCase_ : str = gather_results_from_each_node(lowerCAmelCase_ , lowerCAmelCase_ , args.sync_timeout ) lowerCAmelCase_ : Union[str, Any] = combine_partial_results(lowerCAmelCase_ ) if args.num_return_sequences > 1: lowerCAmelCase_ : Any = save_dir.joinpath('''pseudolabel_results.json''' ) print(f"""Saving aggregated results at {save_path}, intermediate in {json_save_dir}/""" ) save_json(lowerCAmelCase_ , lowerCAmelCase_ ) return lowerCAmelCase_ : Optional[int] = Path(args.data_dir ).joinpath(args.type_path + '''.target''' ) with open(lowerCAmelCase_ ) as f: lowerCAmelCase_ : Dict = [x.rstrip() for x in f.readlines()][: len(lowerCAmelCase_ )] # Calculate metrics, save metrics, and save _generations.txt lowerCAmelCase_ : List[Any] = '''translation''' in args.task lowerCAmelCase_ : List[Any] = calculate_bleu if calc_bleu else calculate_rouge lowerCAmelCase_ : Dict = '''bleu''' if calc_bleu else '''rouge''' lowerCAmelCase_ : Dict = score_fn(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = len(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = time.time() - start_time lowerCAmelCase_ : int = round(runtime / metrics['''n_obs'''] , 4 ) lowerCAmelCase_ : Optional[int] = num_replicas # TODO(@stas00): add whatever metadata to metrics lowerCAmelCase_ : Any = save_dir.joinpath(f"""{args.type_path}_{metric_name}.json""" ) save_json(lowerCAmelCase_ , lowerCAmelCase_ , indent=lowerCAmelCase_ ) print(lowerCAmelCase_ ) write_txt_file(lowerCAmelCase_ , save_dir.joinpath(f"""{args.type_path}_generations.txt""" ) ) if args.debug: write_txt_file(lowerCAmelCase_ , save_dir.joinpath(f"""{args.type_path}.target""" ) ) else: shutil.rmtree(lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> List: lowerCAmelCase_ : Optional[int] = [] for partial_result in partial_results: records.extend(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x["id"] ) lowerCAmelCase_ : Union[str, Any] = [x['''pred'''] for x in records] return preds def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[Dict[str, List]]: # WAIT FOR lots of .json files lowerCAmelCase_ : Any = time.time() logger.info('''waiting for all nodes to finish''' ) lowerCAmelCase_ : Any = None while (time.time() - start_wait) < timeout: lowerCAmelCase_ : int = list(save_dir.glob('''rank_*.json''' ) ) if len(lowerCAmelCase_ ) < num_replicas: continue try: # make sure all json files are fully saved lowerCAmelCase_ : str = lmap(lowerCAmelCase_ , lowerCAmelCase_ ) return json_data except JSONDecodeError: continue else: raise TimeoutError('''Rank 0 gave up on waiting for other processes''' ) # Unreachable if __name__ == "__main__": # Usage for MT: run_generate()
262
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = JukeboxTokenizer SCREAMING_SNAKE_CASE__ : int = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowercase_ ( self ) -> Union[str, Any]: import torch lowerCAmelCase_ : Union[str, Any] = JukeboxTokenizer.from_pretrained('''openai/jukebox-1b-lyrics''' ) lowerCAmelCase_ : Any = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase_ ( self ) -> List[Any]: import torch lowerCAmelCase_ : Any = JukeboxTokenizer.from_pretrained('''openai/jukebox-5b-lyrics''' ) lowerCAmelCase_ : str = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : Tuple = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
262
1
def lowerCAmelCase ( lowerCAmelCase_ )-> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('''The given input must be positive''' ) # get the generated string sequence lowerCAmelCase_ : Tuple = gray_code_sequence_string(lowerCAmelCase_ ) # # convert them to integers for i in range(len(lowerCAmelCase_ ) ): lowerCAmelCase_ : Optional[int] = int(sequence[i] , 2 ) return sequence def lowerCAmelCase ( lowerCAmelCase_ )-> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCAmelCase_ : Any = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCAmelCase_ : List[Any] = gray_code_sequence_string(bit_count - 1 ) lowerCAmelCase_ : str = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCAmelCase_ : List[str] = '''0''' + smaller_sequence[i] sequence.append(lowerCAmelCase_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCAmelCase_ : Optional[Any] = '''1''' + smaller_sequence[i] sequence.append(lowerCAmelCase_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
262
from __future__ import annotations import requests def lowerCAmelCase ( lowerCAmelCase_ )-> dict: lowerCAmelCase_ : List[Any] = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> list[dict]: lowerCAmelCase_ : List[Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCAmelCase_ : Tuple = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> str: lowerCAmelCase_ : Optional[Any] = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join('''* [{title}]({url})'''.format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
262
1
_UpperCAmelCase : int ="""Tobias Carryer""" from time import time class snake_case__: '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase=int(time() ) ) -> str: # noqa: B008 lowerCAmelCase_ : Any = multiplier lowerCAmelCase_ : Tuple = increment lowerCAmelCase_ : Optional[Any] = modulo lowerCAmelCase_ : Optional[Any] = seed def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. _UpperCAmelCase : Optional[Any] =LinearCongruentialGenerator(166_4525, 10_1390_4223, 2 << 31) while True: print(lcg.next_number())
262
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : List[str] =get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCAmelCase : Optional[int] =25_0004 _UpperCAmelCase : Tuple =25_0020 @require_sentencepiece @require_tokenizers class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = MBartTokenizer SCREAMING_SNAKE_CASE__ : Dict = MBartTokenizerFast SCREAMING_SNAKE_CASE__ : Tuple = True SCREAMING_SNAKE_CASE__ : List[str] = True def lowercase_ ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : str = MBartTokenizer(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[int] = MBartTokenizer(__lowercase , keep_accents=__lowercase ) lowerCAmelCase_ : Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCAmelCase_ : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCAmelCase_ : Dict = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowercase_ ( self ) -> Dict: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase_ : int = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase_ : Optional[int] = self.rust_tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : int = self.tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = tempfile.mkdtemp() lowerCAmelCase_ : Union[str, Any] = tokenizer_r.save_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase_ : str = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Tuple = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : int = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Tuple = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Optional[int] = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Optional[int] = tokenizer_p.save_pretrained(__lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Dict = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) @require_torch @require_sentencepiece @require_tokenizers class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = """facebook/mbart-large-en-ro""" SCREAMING_SNAKE_CASE__ : int = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] SCREAMING_SNAKE_CASE__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] SCREAMING_SNAKE_CASE__ : str = [8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2, EN_CODE] @classmethod def lowercase_ ( cls ) -> Optional[int]: lowerCAmelCase_ : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) lowerCAmelCase_ : Optional[Any] = 1 return cls def lowercase_ ( self ) -> Optional[Any]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 2_5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 2_5_0_0_0_4 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 2_5_0_0_2_0 ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Optional[int] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) def lowercase_ ( self ) -> Any: self.assertIn(__lowercase , self.tokenizer.all_special_ids ) lowerCAmelCase_ : Union[str, Any] = [RO_CODE, 8_8_4, 9_0_1_9, 9_6, 9, 9_1_6, 8_6_7_9_2, 3_6, 1_8_7_4_3, 1_5_5_9_6, 5, 2] lowerCAmelCase_ : Tuple = self.tokenizer.decode(__lowercase , skip_special_tokens=__lowercase ) lowerCAmelCase_ : List[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertNotIn(self.tokenizer.eos_token , __lowercase ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Union[str, Any] = ['''this is gunna be a long sentence ''' * 2_0] assert isinstance(src_text[0] , __lowercase ) lowerCAmelCase_ : str = 1_0 lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , max_length=__lowercase , truncation=__lowercase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __lowercase ) self.assertEqual(len(__lowercase ) , __lowercase ) def lowercase_ ( self ) -> int: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [2_5_0_0_2_6, 2_5_0_0_0_1] ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = tempfile.mkdtemp() lowerCAmelCase_ : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = MBartTokenizer.from_pretrained(__lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowercase ) @require_torch def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Tuple = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowercase , return_tensors='''pt''' ) lowerCAmelCase_ : Tuple = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) lowerCAmelCase_ : int = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 1_4) , batch.input_ids.shape ) self.assertEqual((2, 1_4) , batch.attention_mask.shape ) lowerCAmelCase_ : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[Any] = self.tokenizer(self.src_text , padding=__lowercase , truncation=__lowercase , max_length=3 , return_tensors='''pt''' ) lowerCAmelCase_ : Any = self.tokenizer( text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=1_0 , return_tensors='''pt''' ) lowerCAmelCase_ : int = targets['''input_ids'''] lowerCAmelCase_ : Optional[Any] = shift_tokens_right(__lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(__lowercase ) , { # A, test, EOS, en_XX '''input_ids''': [[6_2, 3_0_3_4, 2, 2_5_0_0_0_4]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 2_5_0_0_0_1, } , )
262
1
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class snake_case__( unittest.TestCase ): '''simple docstring''' @property def lowercase_ ( self ) -> Optional[Any]: torch.manual_seed(0 ) lowerCAmelCase_ : int = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Optional[int] = self.dummy_uncond_unet lowerCAmelCase_ : Optional[Any] = PNDMScheduler() lowerCAmelCase_ : int = PNDMPipeline(unet=__lowercase , scheduler=__lowercase ) pndm.to(__lowercase ) pndm.set_progress_bar_config(disable=__lowercase ) lowerCAmelCase_ : List[str] = torch.manual_seed(0 ) lowerCAmelCase_ : Any = pndm(generator=__lowercase , num_inference_steps=2_0 , output_type='''numpy''' ).images lowerCAmelCase_ : str = torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = pndm(generator=__lowercase , num_inference_steps=2_0 , output_type='''numpy''' , return_dict=__lowercase )[0] lowerCAmelCase_ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowerCAmelCase_ : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = '''google/ddpm-cifar10-32''' lowerCAmelCase_ : Dict = UNetaDModel.from_pretrained(__lowercase ) lowerCAmelCase_ : Dict = PNDMScheduler() lowerCAmelCase_ : Any = PNDMPipeline(unet=__lowercase , scheduler=__lowercase ) pndm.to(__lowercase ) pndm.set_progress_bar_config(disable=__lowercase ) lowerCAmelCase_ : Tuple = torch.manual_seed(0 ) lowerCAmelCase_ : List[Any] = pndm(generator=__lowercase , output_type='''numpy''' ).images lowerCAmelCase_ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowerCAmelCase_ : int = np.array([0.15_64, 0.1_46_45, 0.14_06, 0.1_47_15, 0.1_24_25, 0.1_40_45, 0.1_31_15, 0.1_21_75, 0.1_25] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
262
1
# flake8: noqa # Lint as: python3 _UpperCAmelCase : Dict =[ """VerificationMode""", """Version""", """disable_progress_bar""", """enable_progress_bar""", """is_progress_bar_enabled""", """experimental""", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
262
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : str ={ """facebook/vit-mae-base""": """https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json""", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """vit_mae""" def __init__( self , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1e-12 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase=3 , __lowercase=True , __lowercase=1_6 , __lowercase=5_1_2 , __lowercase=8 , __lowercase=2_0_4_8 , __lowercase=0.75 , __lowercase=False , **__lowercase , ) -> str: super().__init__(**__lowercase ) lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : Union[str, Any] = image_size lowerCAmelCase_ : Optional[int] = patch_size lowerCAmelCase_ : Tuple = num_channels lowerCAmelCase_ : List[str] = qkv_bias lowerCAmelCase_ : List[Any] = decoder_num_attention_heads lowerCAmelCase_ : int = decoder_hidden_size lowerCAmelCase_ : Optional[int] = decoder_num_hidden_layers lowerCAmelCase_ : Tuple = decoder_intermediate_size lowerCAmelCase_ : Tuple = mask_ratio lowerCAmelCase_ : Any = norm_pix_loss
262
1
import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class snake_case__: '''simple docstring''' def __init__( self , __lowercase , __lowercase=1_3 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=9_9 , __lowercase=3_2 , __lowercase=5 , __lowercase=4 , __lowercase=3_7 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_1_2 , __lowercase=1_6 , __lowercase=2 , __lowercase=0.02 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : str = parent lowerCAmelCase_ : Dict = batch_size lowerCAmelCase_ : List[str] = seq_length lowerCAmelCase_ : Optional[int] = is_training lowerCAmelCase_ : Optional[Any] = use_input_mask lowerCAmelCase_ : List[Any] = use_token_type_ids lowerCAmelCase_ : Any = use_labels lowerCAmelCase_ : Optional[int] = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : List[str] = num_attention_heads lowerCAmelCase_ : Any = intermediate_size lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : List[Any] = attention_probs_dropout_prob lowerCAmelCase_ : Tuple = max_position_embeddings lowerCAmelCase_ : str = type_vocab_size lowerCAmelCase_ : Dict = type_sequence_label_size lowerCAmelCase_ : List[Any] = initializer_range lowerCAmelCase_ : List[Any] = num_labels lowerCAmelCase_ : List[Any] = num_choices lowerCAmelCase_ : Tuple = scope def lowercase_ ( self ) -> str: lowerCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ : str = None if self.use_input_mask: lowerCAmelCase_ : int = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : Optional[int] = None if self.use_token_type_ids: lowerCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : Tuple = None lowerCAmelCase_ : List[str] = None if self.use_labels: lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ : Dict = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase_ ( self ) -> Optional[Any]: return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[int] = NystromformerModel(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Dict = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase ) lowerCAmelCase_ : Dict = model(__lowercase , token_type_ids=__lowercase ) lowerCAmelCase_ : Dict = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : List[Any] = NystromformerForMaskedLM(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Optional[int] = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Optional[Any] = NystromformerForQuestionAnswering(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : List[Any] = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , start_positions=__lowercase , end_positions=__lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Optional[Any] = self.num_labels lowerCAmelCase_ : List[str] = NystromformerForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Tuple = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Tuple = self.num_labels lowerCAmelCase_ : Optional[int] = NystromformerForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Optional[int] = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Tuple: lowerCAmelCase_ : Union[str, Any] = self.num_choices lowerCAmelCase_ : List[Any] = NystromformerForMultipleChoice(config=__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Tuple = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : Optional[Any] = model( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() ( ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ) : List[Any] = config_and_inputs lowerCAmelCase_ : Dict = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class snake_case__( UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = ( { """feature-extraction""": NystromformerModel, """fill-mask""": NystromformerForMaskedLM, """question-answering""": NystromformerForQuestionAnswering, """text-classification""": NystromformerForSequenceClassification, """token-classification""": NystromformerForTokenClassification, """zero-shot""": NystromformerForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : int = False SCREAMING_SNAKE_CASE__ : int = False def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Any = NystromformerModelTester(self ) lowerCAmelCase_ : str = ConfigTester(self , config_class=__lowercase , hidden_size=3_7 ) def lowercase_ ( self ) -> List[str]: self.config_tester.run_common_tests() def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def lowercase_ ( self ) -> str: lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCAmelCase_ : Optional[int] = type self.model_tester.create_and_check_model(*__lowercase ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__lowercase ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowercase ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowercase ) def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowercase ) @slow def lowercase_ ( self ) -> Optional[int]: for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Optional[int] = NystromformerModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' @slow def lowercase_ ( self ) -> str: lowerCAmelCase_ : Any = NystromformerModel.from_pretrained('''uw-madison/nystromformer-512''' ) lowerCAmelCase_ : Union[str, Any] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): lowerCAmelCase_ : Tuple = model(__lowercase )[0] lowerCAmelCase_ : Optional[Any] = torch.Size((1, 6, 7_6_8) ) self.assertEqual(output.shape , __lowercase ) lowerCAmelCase_ : Dict = torch.tensor( [[[-0.45_32, -0.09_36, 0.51_37], [-0.26_76, 0.06_28, 0.61_86], [-0.36_29, -0.17_26, 0.47_16]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=1e-4 ) ) @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = '''the [MASK] of Belgium is Brussels''' lowerCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained('''uw-madison/nystromformer-512''' ) lowerCAmelCase_ : Optional[int] = NystromformerForMaskedLM.from_pretrained('''uw-madison/nystromformer-512''' ) lowerCAmelCase_ : List[str] = tokenizer(__lowercase , return_tensors='''pt''' ) with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = model(encoding.input_ids ).logits lowerCAmelCase_ : str = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(__lowercase ) , '''capital''' )
262
def lowerCAmelCase ( lowerCAmelCase_ = 10**9 )-> int: lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : str = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value lowerCAmelCase_ : Any = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f"""{solution() = }""")
262
1
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _UpperCAmelCase : str =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use GLPNImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
import inspect import unittest class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: try: import diffusers # noqa: F401 except ImportError: assert False def lowercase_ ( self ) -> List[str]: import diffusers from diffusers.dependency_versions_table import deps lowerCAmelCase_ : Any = inspect.getmembers(__lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowerCAmelCase_ : Optional[int] = '''k-diffusion''' elif backend == "invisible_watermark": lowerCAmelCase_ : Dict = '''invisible-watermark''' assert backend in deps, f"""{backend} is not in the deps table!"""
262
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def lowercase_ ( self , __lowercase ) -> List[str]: with open(__lowercase , encoding='''utf-8''' ) as input_file: lowerCAmelCase_ : Tuple = re.compile(R'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''' ) lowerCAmelCase_ : Tuple = input_file.read() lowerCAmelCase_ : Any = regexp.search(__lowercase ) return match def lowercase_ ( self , __lowercase ) -> Optional[Any]: with open(__lowercase , encoding='''utf-8''' ) as input_file: lowerCAmelCase_ : Optional[int] = re.compile(R'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''' , re.DOTALL ) lowerCAmelCase_ : Union[str, Any] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowerCAmelCase_ : Optional[int] = regexp.finditer(__lowercase ) lowerCAmelCase_ : Any = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : List[Any] = Path('''./datasets''' ) lowerCAmelCase_ : List[Any] = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(f"""open(...) must use utf-8 encoding in {dataset}""" ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[int] = Path('''./datasets''' ) lowerCAmelCase_ : Any = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(f"""print statement found in {dataset}. Use datasets.logger/logging instead.""" )
262
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCAmelCase : Any =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
1
import argparse import os import re import packaging.version _UpperCAmelCase : List[Any] ="""examples/""" _UpperCAmelCase : Optional[int] ={ """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } _UpperCAmelCase : List[Any] ={ """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } _UpperCAmelCase : List[Any] ="""README.md""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : Optional[int] = f.read() lowerCAmelCase_ , lowerCAmelCase_ : List[str] = REPLACE_PATTERNS[pattern] lowerCAmelCase_ : Union[str, Any] = replace.replace('''VERSION''' , lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = re_pattern.sub(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> Dict: for folder, directories, fnames in os.walk(lowerCAmelCase_ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ , pattern='''examples''' ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_=False )-> Tuple: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if not patch: update_version_in_examples(lowerCAmelCase_ ) def lowerCAmelCase ( )-> Optional[Any]: lowerCAmelCase_ : Dict = '''🤗 Transformers currently provides the following architectures''' lowerCAmelCase_ : List[Any] = '''1. Want to contribute a new model?''' with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : Optional[Any] = f.readlines() # Find the start of the list. lowerCAmelCase_ : List[str] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCAmelCase_ : int = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowerCAmelCase_ : Tuple = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(lowerCAmelCase_ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lowerCAmelCase_ ) def lowerCAmelCase ( )-> Any: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowerCAmelCase_ : List[Any] = f.read() lowerCAmelCase_ : Optional[int] = REPLACE_PATTERNS['''init'''][0].search(lowerCAmelCase_ ).groups()[0] return packaging.version.parse(lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_=False )-> Optional[int]: lowerCAmelCase_ : List[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowerCAmelCase_ : int = default_version.base_version elif patch: lowerCAmelCase_ : List[str] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowerCAmelCase_ : Dict = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowerCAmelCase_ : Any = input(f"""Which version are you releasing? [{default_version}]""" ) if len(lowerCAmelCase_ ) == 0: lowerCAmelCase_ : Optional[Any] = default_version print(f"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase_ , patch=lowerCAmelCase_ ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def lowerCAmelCase ( )-> Optional[int]: lowerCAmelCase_ : List[Any] = get_version() lowerCAmelCase_ : Optional[int] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowerCAmelCase_ : int = current_version.base_version # Check with the user we got that right. lowerCAmelCase_ : Tuple = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(lowerCAmelCase_ ) == 0: lowerCAmelCase_ : int = dev_version print(f"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase_ ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") _UpperCAmelCase : int =parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
262
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
1
from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging _UpperCAmelCase : Dict =logging.get_logger(__name__) class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str SCREAMING_SNAKE_CASE__ : str = None @staticmethod def lowercase_ ( ) -> Optional[int]: raise NotImplementedError def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> Dict: raise NotImplementedError def lowercase_ ( self , __lowercase ) -> List[str]: raise NotImplementedError def lowercase_ ( self ) -> Tuple: if not self.is_available(): raise RuntimeError( f"""You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}.""" ) @classmethod def lowercase_ ( cls ) -> Dict: return f"""`pip install {cls.pip_package or cls.name}`""" class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = """optuna""" @staticmethod def lowercase_ ( ) -> int: return is_optuna_available() def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> List[Any]: return run_hp_search_optuna(__lowercase , __lowercase , __lowercase , **__lowercase ) def lowercase_ ( self , __lowercase ) -> int: return default_hp_space_optuna(__lowercase ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """ray""" SCREAMING_SNAKE_CASE__ : int = """'ray[tune]'""" @staticmethod def lowercase_ ( ) -> Optional[int]: return is_ray_available() def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: return run_hp_search_ray(__lowercase , __lowercase , __lowercase , **__lowercase ) def lowercase_ ( self , __lowercase ) -> Optional[int]: return default_hp_space_ray(__lowercase ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """sigopt""" @staticmethod def lowercase_ ( ) -> Dict: return is_sigopt_available() def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> int: return run_hp_search_sigopt(__lowercase , __lowercase , __lowercase , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[str]: return default_hp_space_sigopt(__lowercase ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = """wandb""" @staticmethod def lowercase_ ( ) -> int: return is_wandb_available() def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> int: return run_hp_search_wandb(__lowercase , __lowercase , __lowercase , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: return default_hp_space_wandb(__lowercase ) _UpperCAmelCase : Tuple ={ HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def lowerCAmelCase ( )-> str: lowerCAmelCase_ : str = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(lowerCAmelCase_ ) > 0: lowerCAmelCase_ : Dict = available_backends[0].name if len(lowerCAmelCase_ ) > 1: logger.info( f"""{len(lowerCAmelCase_ )} hyperparameter search backends available. Using {name} as the default.""" ) return name raise RuntimeError( '''No hyperparameter search backend available.\n''' + '''\n'''.join( f""" - To install {backend.name} run {backend.pip_install()}""" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
262
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def lowerCAmelCase ( lowerCAmelCase_=None )-> List[str]: if subparsers is not None: lowerCAmelCase_ : List[str] = subparsers.add_parser('''test''' ) else: lowerCAmelCase_ : Any = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=lowerCAmelCase_ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase_ ) return parser def lowerCAmelCase ( lowerCAmelCase_ )-> str: lowerCAmelCase_ : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: lowerCAmelCase_ : Union[str, Any] = script_name else: lowerCAmelCase_ : List[str] = f"""--config_file={args.config_file} {script_name}""" lowerCAmelCase_ : Optional[int] = ['''accelerate-launch'''] + test_args.split() lowerCAmelCase_ : str = execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def lowerCAmelCase ( )-> List[str]: lowerCAmelCase_ : Tuple = test_command_parser() lowerCAmelCase_ : str = parser.parse_args() test_command(lowerCAmelCase_ ) if __name__ == "__main__": main()
262
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
1
from collections.abc import Iterable from typing import Generic, TypeVar _UpperCAmelCase : int =TypeVar("""_T""") class snake_case__( Generic[_T] ): '''simple docstring''' def __init__( self , __lowercase = None ) -> None: lowerCAmelCase_ : list[_T] = list(iterable or [] ) lowerCAmelCase_ : list[_T] = [] def __len__( self ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def lowercase_ ( self , __lowercase ) -> None: self._stacka.append(__lowercase ) def lowercase_ ( self ) -> _T: lowerCAmelCase_ : int = self._stacka.pop lowerCAmelCase_ : List[str] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError('''Queue is empty''' ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py _UpperCAmelCase : Optional[Any] ="""src/diffusers""" # Matches is_xxx_available() _UpperCAmelCase : Dict =re.compile(R"""is\_([a-z_]*)_available\(\)""") # Matches from xxx import bla _UpperCAmelCase : str =re.compile(R"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") _UpperCAmelCase : Dict =""" {0} = None """ _UpperCAmelCase : List[str] =""" class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, {1}) @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, {1}) """ _UpperCAmelCase : Optional[int] =""" def {0}(*args, **kwargs): requires_backends({0}, {1}) """ def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : List[Any] = _re_backend.findall(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 0: return None return "_and_".join(lowerCAmelCase_ ) def lowerCAmelCase ( )-> Optional[int]: with open(os.path.join(lowerCAmelCase_ , '''__init__.py''' ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : Any = f.readlines() # Get to the point we do the actual imports for type checking lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : Any = {} # Go through the end of the file while line_index < len(lowerCAmelCase_ ): # If the line contains is_backend_available, we grab all objects associated with the `else` block lowerCAmelCase_ : Optional[int] = find_backend(lines[line_index] ) if backend is not None: while not lines[line_index].startswith('''else:''' ): line_index += 1 line_index += 1 lowerCAmelCase_ : Any = [] # Until we unindent, add backend objects to the list while line_index < len(lowerCAmelCase_ ) and len(lines[line_index] ) > 1: lowerCAmelCase_ : Union[str, Any] = lines[line_index] lowerCAmelCase_ : Optional[Any] = _re_single_line_import.search(lowerCAmelCase_ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 if len(lowerCAmelCase_ ) > 0: lowerCAmelCase_ : List[Any] = objects else: line_index += 1 return backend_specific_objects def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: if name.isupper(): return DUMMY_CONSTANT.format(lowerCAmelCase_ ) elif name.islower(): return DUMMY_FUNCTION.format(lowerCAmelCase_ , lowerCAmelCase_ ) else: return DUMMY_CLASS.format(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_=None )-> Union[str, Any]: if backend_specific_objects is None: lowerCAmelCase_ : Optional[int] = read_init() # For special correspondence backend to module name as used in the function requires_modulename lowerCAmelCase_ : Any = {} for backend, objects in backend_specific_objects.items(): lowerCAmelCase_ : List[str] = '''[''' + ''', '''.join(f"""\"{b}\"""" for b in backend.split('''_and_''' ) ) + ''']''' lowerCAmelCase_ : List[Any] = '''# This file is autogenerated by the command `make fix-copies`, do not edit.\n''' dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(lowerCAmelCase_ , lowerCAmelCase_ ) for o in objects] ) lowerCAmelCase_ : int = dummy_file return dummy_files def lowerCAmelCase ( lowerCAmelCase_=False )-> Any: lowerCAmelCase_ : Tuple = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py lowerCAmelCase_ : Union[str, Any] = {'''torch''': '''pt'''} # Locate actual dummy modules and read their content. lowerCAmelCase_ : Tuple = os.path.join(lowerCAmelCase_ , '''utils''' ) lowerCAmelCase_ : int = { backend: os.path.join(lowerCAmelCase_ , f"""dummy_{short_names.get(lowerCAmelCase_ , lowerCAmelCase_ )}_objects.py""" ) for backend in dummy_files.keys() } lowerCAmelCase_ : Union[str, Any] = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(lowerCAmelCase_ ): with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : Union[str, Any] = f.read() else: lowerCAmelCase_ : List[str] = '''''' for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( f"""Updating diffusers.utils.dummy_{short_names.get(lowerCAmelCase_ , lowerCAmelCase_ )}_objects.py as the main """ '''__init__ has new objects.''' ) with open(dummy_file_paths[backend] , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(dummy_files[backend] ) else: raise ValueError( '''The main __init__ has objects that are not present in ''' f"""diffusers.utils.dummy_{short_names.get(lowerCAmelCase_ , lowerCAmelCase_ )}_objects.py. Run `make fix-copies` """ '''to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : Dict =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Union[str, Any] =parser.parse_args() check_dummies(args.fix_and_overwrite)
262
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
1
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
1
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
1
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : torch.FloatTensor class snake_case__( UpperCAmelCase__, UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , __lowercase = 1_6 , __lowercase = 8_8 , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 0.0 , __lowercase = 3_2 , __lowercase = None , __lowercase = False , __lowercase = None , __lowercase = "geglu" , __lowercase = True , __lowercase = True , ) -> List[str]: super().__init__() lowerCAmelCase_ : str = num_attention_heads lowerCAmelCase_ : Tuple = attention_head_dim lowerCAmelCase_ : Optional[int] = num_attention_heads * attention_head_dim lowerCAmelCase_ : Any = in_channels lowerCAmelCase_ : Dict = torch.nn.GroupNorm(num_groups=__lowercase , num_channels=__lowercase , eps=1e-6 , affine=__lowercase ) lowerCAmelCase_ : int = nn.Linear(__lowercase , __lowercase ) # 3. Define transformers blocks lowerCAmelCase_ : Union[str, Any] = nn.ModuleList( [ BasicTransformerBlock( __lowercase , __lowercase , __lowercase , dropout=__lowercase , cross_attention_dim=__lowercase , activation_fn=__lowercase , attention_bias=__lowercase , double_self_attention=__lowercase , norm_elementwise_affine=__lowercase , ) for d in range(__lowercase ) ] ) lowerCAmelCase_ : Union[str, Any] = nn.Linear(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=1 , __lowercase=None , __lowercase = True , ) -> Union[str, Any]: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = hidden_states.shape lowerCAmelCase_ : List[Any] = batch_frames // num_frames lowerCAmelCase_ : Optional[int] = hidden_states lowerCAmelCase_ : Optional[Any] = hidden_states[None, :].reshape(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = hidden_states.permute(0 , 2 , 1 , 3 , 4 ) lowerCAmelCase_ : List[Any] = self.norm(__lowercase ) lowerCAmelCase_ : Dict = hidden_states.permute(0 , 3 , 4 , 2 , 1 ).reshape(batch_size * height * width , __lowercase , __lowercase ) lowerCAmelCase_ : Optional[Any] = self.proj_in(__lowercase ) # 2. Blocks for block in self.transformer_blocks: lowerCAmelCase_ : Optional[int] = block( __lowercase , encoder_hidden_states=__lowercase , timestep=__lowercase , cross_attention_kwargs=__lowercase , class_labels=__lowercase , ) # 3. Output lowerCAmelCase_ : List[Any] = self.proj_out(__lowercase ) lowerCAmelCase_ : Dict = ( hidden_states[None, None, :] .reshape(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) .permute(0 , 3 , 4 , 1 , 2 ) .contiguous() ) lowerCAmelCase_ : Tuple = hidden_states.reshape(__lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Optional[Any] = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=__lowercase )
262
from __future__ import annotations import math class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : str = size # approximate the overall size of segment tree with given value lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] lowerCAmelCase_ : Optional[int] = [0 for i in range(0 , 4 * size )] # flag for lazy update def lowercase_ ( self , __lowercase ) -> int: return idx * 2 def lowercase_ ( self , __lowercase ) -> int: return idx * 2 + 1 def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> None: if left_element == right_element: lowerCAmelCase_ : Tuple = a[left_element - 1] else: lowerCAmelCase_ : int = (left_element + right_element) // 2 self.build(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase ) self.build(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase ) lowerCAmelCase_ : Any = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> bool: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Union[str, Any] = False if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Any = self.lazy[idx] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCAmelCase_ : Dict = val if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = val lowerCAmelCase_ : List[Any] = val lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : List[str] = True return True lowerCAmelCase_ : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) self.update(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) return True def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> int | float: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Optional[Any] = False if left_element != right_element: lowerCAmelCase_ : List[Any] = self.lazy[idx] lowerCAmelCase_ : Dict = self.lazy[idx] lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCAmelCase_ : List[Any] = (left_element + right_element) // 2 lowerCAmelCase_ : Tuple = self.query(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.query(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase ) return max(__lowercase , __lowercase ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , __lowercase , __lowercase ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _UpperCAmelCase : str =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _UpperCAmelCase : List[str] =15 _UpperCAmelCase : Any =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
262
1
import requests def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> None: lowerCAmelCase_ : Dict = {'''Content-Type''': '''application/json'''} lowerCAmelCase_ : int = requests.post(lowerCAmelCase_ , json={'''text''': message_body} , headers=lowerCAmelCase_ ) if response.status_code != 200: lowerCAmelCase_ : List[str] = ( '''Request to slack returned an error ''' f"""{response.status_code}, the response is:\n{response.text}""" ) raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message("""<YOUR MESSAGE BODY>""", """<SLACK CHANNEL URL>""")
262
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py _UpperCAmelCase : Optional[int] ="""src/transformers""" _UpperCAmelCase : str ="""docs/source/en""" _UpperCAmelCase : Optional[int] =""".""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : int = f.readlines() # Find the start prompt. lowerCAmelCase_ : List[Any] = 0 while not lines[start_index].startswith(lowerCAmelCase_ ): start_index += 1 start_index += 1 lowerCAmelCase_ : List[str] = start_index while not lines[end_index].startswith(lowerCAmelCase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | _UpperCAmelCase : Optional[Any] ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. _UpperCAmelCase : Optional[int] =re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _UpperCAmelCase : Dict =re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _UpperCAmelCase : Optional[Any] =re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : Optional[int] =direct_transformers_import(TRANSFORMERS_PATH) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : str = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' , lowerCAmelCase_ ) return [m.group(0 ) for m in matches] def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Tuple = 2 if text == '''✅''' or text == '''❌''' else len(lowerCAmelCase_ ) lowerCAmelCase_ : int = (width - text_length) // 2 lowerCAmelCase_ : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def lowerCAmelCase ( )-> str: lowerCAmelCase_ : Any = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowerCAmelCase_ : Dict = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } lowerCAmelCase_ : List[Any] = {name: config.replace('''Config''' , '''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. lowerCAmelCase_ : Tuple = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = None if attr_name.endswith('''Tokenizer''' ): lowerCAmelCase_ : Union[str, Any] = slow_tokenizers lowerCAmelCase_ : List[str] = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): lowerCAmelCase_ : int = fast_tokenizers lowerCAmelCase_ : Union[str, Any] = attr_name[:-13] elif _re_tf_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = tf_models lowerCAmelCase_ : str = _re_tf_models.match(lowerCAmelCase_ ).groups()[0] elif _re_flax_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = flax_models lowerCAmelCase_ : Union[str, Any] = _re_flax_models.match(lowerCAmelCase_ ).groups()[0] elif _re_pt_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Any = pt_models lowerCAmelCase_ : List[Any] = _re_pt_models.match(lowerCAmelCase_ ).groups()[0] if lookup_dict is not None: while len(lowerCAmelCase_ ) > 0: if attr_name in model_name_to_prefix.values(): lowerCAmelCase_ : Union[str, Any] = True break # Try again after removing the last word in the name lowerCAmelCase_ : Any = ''''''.join(camel_case_split(lowerCAmelCase_ )[:-1] ) # Let's build that table! lowerCAmelCase_ : int = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) lowerCAmelCase_ : Tuple = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). lowerCAmelCase_ : Union[str, Any] = [len(lowerCAmelCase_ ) + 2 for c in columns] lowerCAmelCase_ : Optional[Any] = max([len(lowerCAmelCase_ ) for name in model_names] ) + 2 # Build the table per se lowerCAmelCase_ : Dict = '''|''' + '''|'''.join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for c, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" lowerCAmelCase_ : List[str] = {True: '''✅''', False: '''❌'''} for name in model_names: lowerCAmelCase_ : List[Any] = model_name_to_prefix[name] lowerCAmelCase_ : Union[str, Any] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for l, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + "|\n" return table def lowerCAmelCase ( lowerCAmelCase_=False )-> Tuple: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = _find_text_in_file( filename=os.path.join(lowerCAmelCase_ , '''index.md''' ) , start_prompt='''<!--This table is updated automatically from the auto modules''' , end_prompt='''<!-- End table-->''' , ) lowerCAmelCase_ : Tuple = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCAmelCase_ , '''index.md''' ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Tuple =parser.parse_args() check_model_table(args.fix_and_overwrite)
262
1
import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("""0.12.2"""): raise Exception("""requires fairseq >= 0.12.2""") if version.parse(fairseq.__version__) > version.parse("""2"""): raise Exception("""requires fairseq < v2""") logging.set_verbosity_info() _UpperCAmelCase : Optional[int] =logging.get_logger(__name__) _UpperCAmelCase : int ="""Hello, World!""" _UpperCAmelCase : Dict ="""en_XX""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : str = Path('''data_bin''' ) lowerCAmelCase_ : List[str] = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(lowerCAmelCase_ ).parent ) , checkpoint_file=Path(lowerCAmelCase_ ).name , _name='''xmod_base''' , arch='''xmod_base''' , task='''multilingual_masked_lm''' , data_name_or_path=str(lowerCAmelCase_ ) , bpe='''sentencepiece''' , sentencepiece_model=str(Path(lowerCAmelCase_ ).parent / '''sentencepiece.bpe.model''' ) , src_dict=str(data_dir / '''dict.txt''' ) , ) xmod.eval() # disable dropout print(lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = xmod.model.encoder.sentence_encoder lowerCAmelCase_ : Tuple = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1e-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , '''bottleneck''' , 2 ) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: lowerCAmelCase_ : Union[str, Any] = xmod.model.classification_heads['''mnli'''].out_proj.weight.shape[0] print('''Our X-MOD config:''' , lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = XmodForSequenceClassification(lowerCAmelCase_ ) if classification_head else XmodForMaskedLM(lowerCAmelCase_ ) model.eval() # Now let's copy all the weights. # Embeddings lowerCAmelCase_ : Tuple = xmod_sent_encoder.embed_tokens.weight lowerCAmelCase_ : Dict = xmod_sent_encoder.embed_positions.weight lowerCAmelCase_ : Optional[Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. lowerCAmelCase_ : Optional[int] = xmod_sent_encoder.layernorm_embedding.weight lowerCAmelCase_ : int = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer lowerCAmelCase_ : Optional[int] = model.roberta.encoder.layer[i] lowerCAmelCase_ : Dict = xmod_sent_encoder.layers[i] # self attention lowerCAmelCase_ : List[str] = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError('''Dimensions of self-attention weights do not match.''' ) lowerCAmelCase_ : List[str] = xmod_layer.self_attn.q_proj.weight lowerCAmelCase_ : int = xmod_layer.self_attn.q_proj.bias lowerCAmelCase_ : Union[str, Any] = xmod_layer.self_attn.k_proj.weight lowerCAmelCase_ : int = xmod_layer.self_attn.k_proj.bias lowerCAmelCase_ : List[str] = xmod_layer.self_attn.v_proj.weight lowerCAmelCase_ : Tuple = xmod_layer.self_attn.v_proj.bias # self-attention output lowerCAmelCase_ : Tuple = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError('''Dimensions of self-attention output weights do not match.''' ) lowerCAmelCase_ : Any = xmod_layer.self_attn.out_proj.weight lowerCAmelCase_ : Optional[Any] = xmod_layer.self_attn.out_proj.bias lowerCAmelCase_ : List[str] = xmod_layer.self_attn_layer_norm.weight lowerCAmelCase_ : List[Any] = xmod_layer.self_attn_layer_norm.bias # intermediate lowerCAmelCase_ : Union[str, Any] = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of intermediate weights do not match.''' ) lowerCAmelCase_ : Any = xmod_layer.fca.weight lowerCAmelCase_ : Any = xmod_layer.fca.bias # output lowerCAmelCase_ : str = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of feed-forward weights do not match.''' ) lowerCAmelCase_ : Union[str, Any] = xmod_layer.fca.weight lowerCAmelCase_ : Tuple = xmod_layer.fca.bias lowerCAmelCase_ : Dict = xmod_layer.final_layer_norm.weight lowerCAmelCase_ : Dict = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: lowerCAmelCase_ : Optional[Any] = xmod_layer.adapter_layer_norm.weight lowerCAmelCase_ : Dict = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError('''Lists of language adapters do not match.''' ) for lang_code, adapter in xmod_layer.adapter_modules.items(): lowerCAmelCase_ : Optional[int] = bert_output.adapter_modules[lang_code] lowerCAmelCase_ : Any = xmod_layer.adapter_modules[lang_code] lowerCAmelCase_ : int = from_adapter.fca.weight lowerCAmelCase_ : str = from_adapter.fca.bias lowerCAmelCase_ : Any = from_adapter.fca.weight lowerCAmelCase_ : Dict = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: lowerCAmelCase_ : str = xmod_sent_encoder.layer_norm.weight lowerCAmelCase_ : str = xmod_sent_encoder.layer_norm.bias if classification_head: lowerCAmelCase_ : Optional[int] = xmod.model.classification_heads['''mnli'''].dense.weight lowerCAmelCase_ : str = xmod.model.classification_heads['''mnli'''].dense.bias lowerCAmelCase_ : Any = xmod.model.classification_heads['''mnli'''].out_proj.weight lowerCAmelCase_ : Union[str, Any] = xmod.model.classification_heads['''mnli'''].out_proj.bias else: # LM Head lowerCAmelCase_ : Tuple = xmod.model.encoder.lm_head.dense.weight lowerCAmelCase_ : int = xmod.model.encoder.lm_head.dense.bias lowerCAmelCase_ : Dict = xmod.model.encoder.lm_head.layer_norm.weight lowerCAmelCase_ : int = xmod.model.encoder.lm_head.layer_norm.bias lowerCAmelCase_ : List[str] = xmod.model.encoder.lm_head.weight lowerCAmelCase_ : Tuple = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. lowerCAmelCase_ : Dict = xmod.encode(lowerCAmelCase_ ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = model(lowerCAmelCase_ )[0] if classification_head: lowerCAmelCase_ : List[Any] = xmod.model.classification_heads['''mnli'''](xmod.extract_features(lowerCAmelCase_ ) ) else: lowerCAmelCase_ : Any = xmod.model(lowerCAmelCase_ , lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape , their_output.shape ) lowerCAmelCase_ : Any = torch.max(torch.abs(our_output - their_output ) ).item() print(f"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 lowerCAmelCase_ : List[str] = torch.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-3 ) print('''Do both models output the same tensors?''' , '''🔥''' if success else '''💩''' ) if not success: raise Exception('''Something went wRoNg''' ) Path(lowerCAmelCase_ ).mkdir(parents=lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] =argparse.ArgumentParser() # Required parameters parser.add_argument( """--xmod_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--classification_head""", action="""store_true""", help="""Whether to convert a final classification head.""" ) _UpperCAmelCase : str =parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
262
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
1
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Any =datasets.logging.get_logger(__name__) _UpperCAmelCase : Any ="""\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\", author = \"Moosavi, Nafise Sadat and Strube, Michael\", booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\", month = aug, year = \"2016\", address = \"Berlin, Germany\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/P16-1060\", doi = \"10.18653/v1/P16-1060\", pages = \"632--642\", } """ _UpperCAmelCase : str ="""\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. """ _UpperCAmelCase : str =""" Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting 'keep_singletons=False', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs. min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: 'mentions': mentions 'muc': MUC metric [Vilain et al, 1995] 'bcub': B-cubed [Bagga and Baldwin, 1998] 'ceafe': CEAFe [Luo et al., 2005] 'lea': LEA [Moosavi and Strube, 2016] 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric('coval') >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -', ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)', ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)', ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -', ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -', ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {'mentions/recall': 1.0,[...] 'conll_score': 100.0} """ def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_="dummy_doc" )-> Optional[int]: lowerCAmelCase_ : List[str] = {doc: key_lines} lowerCAmelCase_ : str = {doc: sys_lines} lowerCAmelCase_ : Optional[Any] = {} lowerCAmelCase_ : str = 0 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : int = 0 lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : int = 0 lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = reader.get_doc_mentions(lowerCAmelCase_ , key_doc_lines[doc] , lowerCAmelCase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowerCAmelCase_ : Any = reader.set_annotated_parse_trees(lowerCAmelCase_ , key_doc_lines[doc] , lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = reader.get_doc_mentions(lowerCAmelCase_ , sys_doc_lines[doc] , lowerCAmelCase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowerCAmelCase_ : Union[str, Any] = reader.set_annotated_parse_trees(lowerCAmelCase_ , key_doc_lines[doc] , lowerCAmelCase_ , lowerCAmelCase_ ) if remove_nested: lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = reader.remove_nested_coref_mentions(lowerCAmelCase_ , lowerCAmelCase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = reader.remove_nested_coref_mentions(lowerCAmelCase_ , lowerCAmelCase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowerCAmelCase_ : int = reader.get_mention_assignments(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = reader.get_mention_assignments(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Union[str, Any] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: lowerCAmelCase_ : List[Any] = get_coref_infos(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Union[str, Any] = {} lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : List[Any] = 0 for name, metric in metrics: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = evaluator.evaluate_documents(lowerCAmelCase_ , lowerCAmelCase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , f"""Recall: {recall * 100:.2f}""" , f""" Precision: {precision * 100:.2f}""" , f""" F1: {fa * 100:.2f}""" , ) if conll_subparts_num == 3: lowerCAmelCase_ : int = (conll / 3) * 100 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: lowerCAmelCase_ : Optional[Any] = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowerCAmelCase_ : List[Any] = line.split()[5] if not parse_col == "-": lowerCAmelCase_ : Union[str, Any] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case__( datasets.Metric ): '''simple docstring''' def lowercase_ ( self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase=True , __lowercase=False , __lowercase=False , __lowercase=False ) -> Optional[Any]: lowerCAmelCase_ : Dict = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowerCAmelCase_ : Tuple = util.check_gold_parse_annotation(__lowercase ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowerCAmelCase_ : List[str] = evaluate( key_lines=__lowercase , sys_lines=__lowercase , metrics=__lowercase , NP_only=__lowercase , remove_nested=__lowercase , keep_singletons=__lowercase , min_span=__lowercase , ) return score
262
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : Tuple =None _UpperCAmelCase : int =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Any ={ """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """facebook/nllb-large-en-ro""": 1024, """facebook/nllb-200-distilled-600M""": 1024, } # fmt: off _UpperCAmelCase : Any =["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : int = NllbTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=False , **__lowercase , ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : int = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : List[Any] = legacy_behaviour super().__init__( vocab_file=__lowercase , tokenizer_file=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , additional_special_tokens=__lowercase , legacy_behaviour=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_file lowerCAmelCase_ : List[Any] = False if not self.vocab_file else True lowerCAmelCase_ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) lowerCAmelCase_ : Optional[Any] = { lang_code: self.convert_tokens_to_ids(__lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase_ : Any = src_lang if src_lang is not None else '''eng_Latn''' lowerCAmelCase_ : str = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase_ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowercase_ ( self ) -> str: return self._src_lang @src_lang.setter def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : int = self(__lowercase , add_special_tokens=__lowercase , return_tensors=__lowercase , **__lowercase ) lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) lowerCAmelCase_ : List[Any] = tgt_lang_id return inputs def lowercase_ ( self , __lowercase , __lowercase = "eng_Latn" , __lowercase = None , __lowercase = "fra_Latn" , **__lowercase , ) -> BatchEncoding: lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : List[str] = tgt_lang return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def lowercase_ ( self ) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[str] = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[str] = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Optional[int] = [self.cur_lang_code] lowerCAmelCase_ : List[Any] = [self.eos_token_id] lowerCAmelCase_ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Any = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Any = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Any = [self.cur_lang_code] lowerCAmelCase_ : Any = [self.eos_token_id] lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ): copyfile(self.vocab_file , __lowercase ) return (out_vocab_file,)
262
1
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _UpperCAmelCase : Optional[Any] =NewType("""DataClass""", Any) _UpperCAmelCase : Dict =NewType("""DataClassType""", Any) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase ( lowerCAmelCase_ )-> Callable[[str], Any]: lowerCAmelCase_ : str = {str(lowerCAmelCase_ ): choice for choice in choices} return lambda lowerCAmelCase_ : str_to_choice.get(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( *, lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = None , **lowerCAmelCase_ , )-> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase_ : Dict = {} if aliases is not None: lowerCAmelCase_ : str = aliases if help is not None: lowerCAmelCase_ : Tuple = help return dataclasses.field(metadata=lowerCAmelCase_ , default=lowerCAmelCase_ , default_factory=lowerCAmelCase_ , **lowerCAmelCase_ ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Iterable[DataClassType] def __init__( self , __lowercase , **__lowercase ) -> List[str]: # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase_ : Optional[int] = ArgumentDefaultsHelpFormatter super().__init__(**__lowercase ) if dataclasses.is_dataclass(__lowercase ): lowerCAmelCase_ : Union[str, Any] = [dataclass_types] lowerCAmelCase_ : List[Any] = list(__lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowercase ) @staticmethod def lowercase_ ( __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[Any] = f"""--{field.name}""" lowerCAmelCase_ : Tuple = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowercase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase_ : List[str] = kwargs.pop('''aliases''' , [] ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[Any] = [aliases] lowerCAmelCase_ : Any = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(__lowercase , '''UnionType''' ) and isinstance(__lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowercase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(__lowercase ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase_ : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase_ : Dict = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase_ : str = ( field.type.__args__[0] if isinstance(__lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase_ : List[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase_ : Dict = {} if origin_type is Literal or (isinstance(field.type , __lowercase ) and issubclass(field.type , __lowercase )): if origin_type is Literal: lowerCAmelCase_ : Optional[Any] = field.type.__args__ else: lowerCAmelCase_ : int = [x.value for x in field.type] lowerCAmelCase_ : str = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase_ : str = field.default else: lowerCAmelCase_ : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase_ : Tuple = copy(__lowercase ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase_ : Dict = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase_ : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase_ : int = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase_ : List[Any] = True elif isclass(__lowercase ) and issubclass(__lowercase , __lowercase ): lowerCAmelCase_ : Union[str, Any] = field.type.__args__[0] lowerCAmelCase_ : Dict = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : Any = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase_ : Dict = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : List[Any] = field.default_factory() else: lowerCAmelCase_ : int = True parser.add_argument(__lowercase , *__lowercase , **__lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase_ : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: if hasattr(__lowercase , '''_argument_group_name''' ): lowerCAmelCase_ : str = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase_ : Dict = self try: lowerCAmelCase_ : Dict[str, type] = get_type_hints(__lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(__lowercase ): lowerCAmelCase_ : Any = '''.'''.join(map(__lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(__lowercase ): if not field.init: continue lowerCAmelCase_ : Optional[int] = type_hints[field.name] self._parse_dataclass_field(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase=None , __lowercase=False , __lowercase=True , __lowercase=None , __lowercase=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase_ : str = [] if args_filename: args_files.append(Path(__lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase_ : str = ArgumentParser() args_file_parser.add_argument(__lowercase , type=__lowercase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = args_file_parser.parse_known_args(args=__lowercase ) lowerCAmelCase_ : int = vars(__lowercase ).get(args_file_flag.lstrip('''-''' ) , __lowercase ) if cmd_args_file_paths: args_files.extend([Path(__lowercase ) for p in cmd_args_file_paths] ) lowerCAmelCase_ : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.parse_known_args(args=__lowercase ) lowerCAmelCase_ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase_ : str = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : str = {k: v for k, v in vars(__lowercase ).items() if k in keys} for k in keys: delattr(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = dtype(**__lowercase ) outputs.append(__lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : int = set(args.keys() ) lowerCAmelCase_ : str = [] for dtype in self.dataclass_types: lowerCAmelCase_ : int = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : List[str] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase_ : List[str] = dtype(**__lowercase ) outputs.append(__lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(__lowercase )}""" ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: with open(Path(__lowercase ) , encoding='''utf-8''' ) as open_json_file: lowerCAmelCase_ : Dict = json.loads(open_json_file.read() ) lowerCAmelCase_ : str = self.parse_dict(__lowercase , allow_extra_keys=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : Optional[Any] = self.parse_dict(yaml.safe_load(Path(__lowercase ).read_text() ) , allow_extra_keys=__lowercase ) return tuple(__lowercase )
262
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCAmelCase : List[Any] ={ """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] =[ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _UpperCAmelCase : Union[str, Any] =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
262
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: return EnvironmentCommand() class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @staticmethod def lowercase_ ( __lowercase ) -> List[Any]: lowerCAmelCase_ : List[str] = parser.add_parser('''env''' ) download_parser.set_defaults(func=__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = huggingface_hub.__version__ lowerCAmelCase_ : str = '''not installed''' lowerCAmelCase_ : str = '''NA''' if is_torch_available(): import torch lowerCAmelCase_ : Any = torch.__version__ lowerCAmelCase_ : str = torch.cuda.is_available() lowerCAmelCase_ : List[str] = '''not installed''' if is_transformers_available(): import transformers lowerCAmelCase_ : Any = transformers.__version__ lowerCAmelCase_ : Optional[Any] = '''not installed''' if is_accelerate_available(): import accelerate lowerCAmelCase_ : List[Any] = accelerate.__version__ lowerCAmelCase_ : List[str] = '''not installed''' if is_xformers_available(): import xformers lowerCAmelCase_ : Optional[Any] = xformers.__version__ lowerCAmelCase_ : int = { '''`diffusers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Huggingface_hub version''': hub_version, '''Transformers version''': transformers_version, '''Accelerate version''': accelerate_version, '''xFormers version''': xformers_version, '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def lowercase_ ( __lowercase ) -> str: return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
262
1
import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : List[str] =logging.get_logger(__name__) def lowerCAmelCase ( lowerCAmelCase_ )-> Dict: print('''Loading config file...''' ) def flatten_yaml_as_dict(lowerCAmelCase_ , lowerCAmelCase_="" , lowerCAmelCase_="." ): lowerCAmelCase_ : Tuple = [] for k, v in d.items(): lowerCAmelCase_ : Union[str, Any] = parent_key + sep + k if parent_key else k if isinstance(lowerCAmelCase_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(lowerCAmelCase_ , lowerCAmelCase_ , sep=lowerCAmelCase_ ).items() ) else: items.append((new_key, v) ) return dict(lowerCAmelCase_ ) lowerCAmelCase_ : Dict = argparse.Namespace() with open(lowerCAmelCase_ , '''r''' ) as yaml_file: try: lowerCAmelCase_ : Optional[int] = yaml.load(lowerCAmelCase_ , Loader=yaml.FullLoader ) lowerCAmelCase_ : Any = flatten_yaml_as_dict(lowerCAmelCase_ ) for k, v in flat_cfg.items(): setattr(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) except yaml.YAMLError as exc: logger.error('''Error while loading config file: {}. Error message: {}'''.format(lowerCAmelCase_ , str(lowerCAmelCase_ ) ) ) return config def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: lowerCAmelCase_ : Any = MobileViTVaConfig() lowerCAmelCase_ : Optional[Any] = False # dataset if task_name.startswith('''imagenet1k_''' ): lowerCAmelCase_ : Union[str, Any] = 1_000 if int(task_name.strip().split('''_''' )[-1] ) == 384: lowerCAmelCase_ : List[str] = 384 else: lowerCAmelCase_ : List[str] = 256 lowerCAmelCase_ : List[str] = '''imagenet-1k-id2label.json''' elif task_name.startswith('''imagenet21k_to_1k_''' ): lowerCAmelCase_ : List[str] = 21_000 if int(task_name.strip().split('''_''' )[-1] ) == 384: lowerCAmelCase_ : int = 384 else: lowerCAmelCase_ : int = 256 lowerCAmelCase_ : Union[str, Any] = '''imagenet-22k-id2label.json''' elif task_name.startswith('''ade20k_''' ): lowerCAmelCase_ : Optional[Any] = 151 lowerCAmelCase_ : Optional[int] = 512 lowerCAmelCase_ : int = '''ade20k-id2label.json''' lowerCAmelCase_ : Dict = True elif task_name.startswith('''voc_''' ): lowerCAmelCase_ : Tuple = 21 lowerCAmelCase_ : Dict = 512 lowerCAmelCase_ : Optional[int] = '''pascal-voc-id2label.json''' lowerCAmelCase_ : Optional[int] = True # orig_config lowerCAmelCase_ : Union[str, Any] = load_orig_config_file(lowerCAmelCase_ ) assert getattr(lowerCAmelCase_ , '''model.classification.name''' , -1 ) == "mobilevit_v2", "Invalid model" lowerCAmelCase_ : Any = getattr(lowerCAmelCase_ , '''model.classification.mitv2.width_multiplier''' , 1.0 ) assert ( getattr(lowerCAmelCase_ , '''model.classification.mitv2.attn_norm_layer''' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" lowerCAmelCase_ : Dict = getattr(lowerCAmelCase_ , '''model.classification.activation.name''' , '''swish''' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: lowerCAmelCase_ : Optional[int] = getattr(lowerCAmelCase_ , '''model.segmentation.output_stride''' , 16 ) if "_deeplabv3" in task_name: lowerCAmelCase_ : Optional[Any] = getattr(lowerCAmelCase_ , '''model.segmentation.deeplabv3.aspp_rates''' , [12, 24, 36] ) lowerCAmelCase_ : Dict = getattr(lowerCAmelCase_ , '''model.segmentation.deeplabv3.aspp_out_channels''' , 512 ) lowerCAmelCase_ : Any = getattr(lowerCAmelCase_ , '''model.segmentation.deeplabv3.aspp_dropout''' , 0.1 ) # id2label lowerCAmelCase_ : int = '''huggingface/label-files''' lowerCAmelCase_ : Optional[Any] = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase_ : List[str] = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} lowerCAmelCase_ : Dict = idalabel lowerCAmelCase_ : int = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Optional[int] = dct.pop(lowerCAmelCase_ ) lowerCAmelCase_ : Union[str, Any] = val def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_=False )-> Union[str, Any]: if base_model: lowerCAmelCase_ : List[str] = '''''' else: lowerCAmelCase_ : Dict = '''mobilevitv2.''' lowerCAmelCase_ : str = [] for k in state_dict.keys(): if k[:8] == "encoder.": lowerCAmelCase_ : List[Any] = k[8:] else: lowerCAmelCase_ : List[Any] = k if ".block." in k: lowerCAmelCase_ : Optional[int] = k_new.replace('''.block.''' , '''.''' ) if ".conv." in k: lowerCAmelCase_ : int = k_new.replace('''.conv.''' , '''.convolution.''' ) if ".norm." in k: lowerCAmelCase_ : str = k_new.replace('''.norm.''' , '''.normalization.''' ) if "conv_1." in k: lowerCAmelCase_ : List[str] = k_new.replace('''conv_1.''' , f"""{model_prefix}conv_stem.""" ) for i in [1, 2]: if f"""layer_{i}.""" in k: lowerCAmelCase_ : Optional[int] = k_new.replace(f"""layer_{i}.""" , f"""{model_prefix}encoder.layer.{i-1}.layer.""" ) if ".exp_1x1." in k: lowerCAmelCase_ : int = k_new.replace('''.exp_1x1.''' , '''.expand_1x1.''' ) if ".red_1x1." in k: lowerCAmelCase_ : List[Any] = k_new.replace('''.red_1x1.''' , '''.reduce_1x1.''' ) for i in [3, 4, 5]: if f"""layer_{i}.0.""" in k: lowerCAmelCase_ : Tuple = k_new.replace(f"""layer_{i}.0.""" , f"""{model_prefix}encoder.layer.{i-1}.downsampling_layer.""" ) if f"""layer_{i}.1.local_rep.0.""" in k: lowerCAmelCase_ : Dict = k_new.replace(f"""layer_{i}.1.local_rep.0.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_kxk.""" ) if f"""layer_{i}.1.local_rep.1.""" in k: lowerCAmelCase_ : str = k_new.replace(f"""layer_{i}.1.local_rep.1.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_1x1.""" ) for i in [3, 4, 5]: if i == 3: lowerCAmelCase_ : Union[str, Any] = [0, 1] elif i == 4: lowerCAmelCase_ : Tuple = [0, 1, 2, 3] elif i == 5: lowerCAmelCase_ : Tuple = [0, 1, 2] for j in j_in: if f"""layer_{i}.1.global_rep.{j}.""" in k: lowerCAmelCase_ : str = k_new.replace( f"""layer_{i}.1.global_rep.{j}.""" , f"""{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.""" ) if f"""layer_{i}.1.global_rep.{j+1}.""" in k: lowerCAmelCase_ : str = k_new.replace( f"""layer_{i}.1.global_rep.{j+1}.""" , f"""{model_prefix}encoder.layer.{i-1}.layernorm.""" ) if f"""layer_{i}.1.conv_proj.""" in k: lowerCAmelCase_ : Optional[Any] = k_new.replace(f"""layer_{i}.1.conv_proj.""" , f"""{model_prefix}encoder.layer.{i-1}.conv_projection.""" ) if "pre_norm_attn.0." in k: lowerCAmelCase_ : Union[str, Any] = k_new.replace('''pre_norm_attn.0.''' , '''layernorm_before.''' ) if "pre_norm_attn.1." in k: lowerCAmelCase_ : Dict = k_new.replace('''pre_norm_attn.1.''' , '''attention.''' ) if "pre_norm_ffn.0." in k: lowerCAmelCase_ : Dict = k_new.replace('''pre_norm_ffn.0.''' , '''layernorm_after.''' ) if "pre_norm_ffn.1." in k: lowerCAmelCase_ : Optional[Any] = k_new.replace('''pre_norm_ffn.1.''' , '''ffn.conv1.''' ) if "pre_norm_ffn.3." in k: lowerCAmelCase_ : List[str] = k_new.replace('''pre_norm_ffn.3.''' , '''ffn.conv2.''' ) if "classifier.1." in k: lowerCAmelCase_ : Tuple = k_new.replace('''classifier.1.''' , '''classifier.''' ) if "seg_head." in k: lowerCAmelCase_ : Any = k_new.replace('''seg_head.''' , '''segmentation_head.''' ) if ".aspp_layer." in k: lowerCAmelCase_ : Optional[Any] = k_new.replace('''.aspp_layer.''' , '''.''' ) if ".aspp_pool." in k: lowerCAmelCase_ : int = k_new.replace('''.aspp_pool.''' , '''.''' ) rename_keys.append((k, k_new) ) return rename_keys def lowerCAmelCase ( lowerCAmelCase_ )-> int: lowerCAmelCase_ : Any = [] for k in state_dict.keys(): if k.startswith('''seg_head.aux_head.''' ): keys_to_ignore.append(lowerCAmelCase_ ) for k in keys_to_ignore: state_dict.pop(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( )-> int: lowerCAmelCase_ : Optional[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" lowerCAmelCase_ : Dict = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : Dict = get_mobilevitva_config(lowerCAmelCase_ , lowerCAmelCase_ ) # load original state_dict lowerCAmelCase_ : Dict = torch.load(lowerCAmelCase_ , map_location='''cpu''' ) # load huggingface model if task_name.startswith('''ade20k_''' ) or task_name.startswith('''voc_''' ): lowerCAmelCase_ : Dict = MobileViTVaForSemanticSegmentation(lowerCAmelCase_ ).eval() lowerCAmelCase_ : Tuple = False else: lowerCAmelCase_ : Tuple = MobileViTVaForImageClassification(lowerCAmelCase_ ).eval() lowerCAmelCase_ : List[str] = False # remove and rename some keys of load the original model lowerCAmelCase_ : Optional[int] = checkpoint remove_unused_keys(lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = create_rename_keys(lowerCAmelCase_ , base_model=lowerCAmelCase_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # load modified state_dict model.load_state_dict(lowerCAmelCase_ ) # Check outputs on an image, prepared by MobileViTImageProcessor lowerCAmelCase_ : List[str] = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) lowerCAmelCase_ : str = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowerCAmelCase_ : Dict = model(**lowerCAmelCase_ ) # verify classification model if task_name.startswith('''imagenet''' ): lowerCAmelCase_ : str = outputs.logits lowerCAmelCase_ : List[str] = logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('''imagenet1k_256''' ) and config.width_multiplier == 1.0: # expected_logits for base variant lowerCAmelCase_ : List[Any] = torch.tensor([-1.6_3_3_6e0_0, -7.3_2_0_4e-0_2, -5.1_8_8_3e-0_1] ) assert torch.allclose(logits[0, :3] , lowerCAmelCase_ , atol=1e-4 ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(f"""Saving model {task_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _UpperCAmelCase : Optional[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you'd like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : Dict =parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
262
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = JukeboxTokenizer SCREAMING_SNAKE_CASE__ : int = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowercase_ ( self ) -> Union[str, Any]: import torch lowerCAmelCase_ : Union[str, Any] = JukeboxTokenizer.from_pretrained('''openai/jukebox-1b-lyrics''' ) lowerCAmelCase_ : Any = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase_ ( self ) -> List[Any]: import torch lowerCAmelCase_ : Any = JukeboxTokenizer.from_pretrained('''openai/jukebox-5b-lyrics''' ) lowerCAmelCase_ : str = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : Tuple = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
262
1
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
from __future__ import annotations import requests def lowerCAmelCase ( lowerCAmelCase_ )-> dict: lowerCAmelCase_ : List[Any] = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> list[dict]: lowerCAmelCase_ : List[Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCAmelCase_ : Tuple = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> str: lowerCAmelCase_ : Optional[Any] = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join('''* [{title}]({url})'''.format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
262
1
def lowerCAmelCase ( lowerCAmelCase_ = 4_000_000 )-> int: lowerCAmelCase_ : Tuple = [0, 1] lowerCAmelCase_ : List[Any] = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i += 1 lowerCAmelCase_ : str = 0 for j in range(len(lowerCAmelCase_ ) - 1 ): if fib[j] % 2 == 0: total += fib[j] return total if __name__ == "__main__": print(f"""{solution() = }""")
262
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : List[str] =get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCAmelCase : Optional[int] =25_0004 _UpperCAmelCase : Tuple =25_0020 @require_sentencepiece @require_tokenizers class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = MBartTokenizer SCREAMING_SNAKE_CASE__ : Dict = MBartTokenizerFast SCREAMING_SNAKE_CASE__ : Tuple = True SCREAMING_SNAKE_CASE__ : List[str] = True def lowercase_ ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : str = MBartTokenizer(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[int] = MBartTokenizer(__lowercase , keep_accents=__lowercase ) lowerCAmelCase_ : Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCAmelCase_ : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCAmelCase_ : Dict = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowercase_ ( self ) -> Dict: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase_ : int = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase_ : Optional[int] = self.rust_tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : int = self.tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = tempfile.mkdtemp() lowerCAmelCase_ : Union[str, Any] = tokenizer_r.save_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase_ : str = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Tuple = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : int = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Tuple = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Optional[int] = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Optional[int] = tokenizer_p.save_pretrained(__lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Dict = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) @require_torch @require_sentencepiece @require_tokenizers class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = """facebook/mbart-large-en-ro""" SCREAMING_SNAKE_CASE__ : int = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] SCREAMING_SNAKE_CASE__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] SCREAMING_SNAKE_CASE__ : str = [8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2, EN_CODE] @classmethod def lowercase_ ( cls ) -> Optional[int]: lowerCAmelCase_ : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) lowerCAmelCase_ : Optional[Any] = 1 return cls def lowercase_ ( self ) -> Optional[Any]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 2_5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 2_5_0_0_0_4 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 2_5_0_0_2_0 ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Optional[int] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) def lowercase_ ( self ) -> Any: self.assertIn(__lowercase , self.tokenizer.all_special_ids ) lowerCAmelCase_ : Union[str, Any] = [RO_CODE, 8_8_4, 9_0_1_9, 9_6, 9, 9_1_6, 8_6_7_9_2, 3_6, 1_8_7_4_3, 1_5_5_9_6, 5, 2] lowerCAmelCase_ : Tuple = self.tokenizer.decode(__lowercase , skip_special_tokens=__lowercase ) lowerCAmelCase_ : List[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertNotIn(self.tokenizer.eos_token , __lowercase ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Union[str, Any] = ['''this is gunna be a long sentence ''' * 2_0] assert isinstance(src_text[0] , __lowercase ) lowerCAmelCase_ : str = 1_0 lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , max_length=__lowercase , truncation=__lowercase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __lowercase ) self.assertEqual(len(__lowercase ) , __lowercase ) def lowercase_ ( self ) -> int: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [2_5_0_0_2_6, 2_5_0_0_0_1] ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = tempfile.mkdtemp() lowerCAmelCase_ : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = MBartTokenizer.from_pretrained(__lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowercase ) @require_torch def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Tuple = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowercase , return_tensors='''pt''' ) lowerCAmelCase_ : Tuple = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) lowerCAmelCase_ : int = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 1_4) , batch.input_ids.shape ) self.assertEqual((2, 1_4) , batch.attention_mask.shape ) lowerCAmelCase_ : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[Any] = self.tokenizer(self.src_text , padding=__lowercase , truncation=__lowercase , max_length=3 , return_tensors='''pt''' ) lowerCAmelCase_ : Any = self.tokenizer( text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=1_0 , return_tensors='''pt''' ) lowerCAmelCase_ : int = targets['''input_ids'''] lowerCAmelCase_ : Optional[Any] = shift_tokens_right(__lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(__lowercase ) , { # A, test, EOS, en_XX '''input_ids''': [[6_2, 3_0_3_4, 2, 2_5_0_0_0_4]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 2_5_0_0_0_1, } , )
262
1
import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self , __lowercase , __lowercase ) -> int: lowerCAmelCase_ : List[str] = jnp.ones((batch_size, length) ) / length return scores def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Dict = None lowerCAmelCase_ : Tuple = 2_0 lowerCAmelCase_ : int = self._get_uniform_logits(batch_size=2 , length=__lowercase ) # tweak scores to not be uniform anymore lowerCAmelCase_ : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch lowerCAmelCase_ : str = scores.at[1, 1_0].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax lowerCAmelCase_ : Optional[int] = jax.nn.softmax(__lowercase , axis=-1 ) lowerCAmelCase_ : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowerCAmelCase_ : Optional[int] = FlaxTemperatureLogitsWarper(temperature=1.3 ) lowerCAmelCase_ : Any = jax.nn.softmax(temp_dist_warper_sharper(__lowercase , scores.copy() , cur_len=__lowercase ) , axis=-1 ) lowerCAmelCase_ : Optional[int] = jax.nn.softmax(temp_dist_warper_smoother(__lowercase , scores.copy() , cur_len=__lowercase ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1e-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1e-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Union[str, Any] = None lowerCAmelCase_ : Dict = 1_0 lowerCAmelCase_ : int = 2 # create ramp distribution lowerCAmelCase_ : Dict = np.broadcast_to(np.arange(__lowercase )[None, :] , (batch_size, vocab_size) ).copy() lowerCAmelCase_ : Dict = ramp_logits[1:, : vocab_size // 2] + vocab_size lowerCAmelCase_ : Dict = FlaxTopKLogitsWarper(3 ) lowerCAmelCase_ : int = top_k_warp(__lowercase , __lowercase , cur_len=__lowercase ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case lowerCAmelCase_ : Optional[Any] = 5 lowerCAmelCase_ : Union[str, Any] = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) lowerCAmelCase_ : Dict = np.broadcast_to(np.arange(__lowercase )[None, :] , (batch_size, length) ).copy() lowerCAmelCase_ : Any = top_k_warp_safety_check(__lowercase , __lowercase , cur_len=__lowercase ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Dict = None lowerCAmelCase_ : int = 1_0 lowerCAmelCase_ : Optional[int] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) lowerCAmelCase_ : Dict = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) lowerCAmelCase_ : int = FlaxTopPLogitsWarper(0.8 ) lowerCAmelCase_ : int = np.exp(top_p_warp(__lowercase , __lowercase , cur_len=__lowercase ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 lowerCAmelCase_ : Union[str, Any] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(__lowercase , __lowercase , atol=1e-3 ) ) # check edge cases with negative and extreme logits lowerCAmelCase_ : Dict = np.broadcast_to(np.arange(__lowercase )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme lowerCAmelCase_ : Any = ramp_logits[1] * 1_00.0 # make sure at least 2 tokens are kept lowerCAmelCase_ : Dict = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) lowerCAmelCase_ : Optional[Any] = top_p_warp(__lowercase , __lowercase , cur_len=__lowercase ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : int = 2_0 lowerCAmelCase_ : List[str] = 4 lowerCAmelCase_ : int = 0 lowerCAmelCase_ : Union[str, Any] = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=__lowercase ) # check that min length is applied at length 5 lowerCAmelCase_ : Any = ids_tensor((batch_size, 2_0) , vocab_size=2_0 ) lowerCAmelCase_ : Tuple = 5 lowerCAmelCase_ : Optional[Any] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : Any = min_dist_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 lowerCAmelCase_ : Optional[Any] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : List[str] = 1_5 lowerCAmelCase_ : Optional[int] = min_dist_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertFalse(jnp.isinf(__lowercase ).any() ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = 2_0 lowerCAmelCase_ : Optional[Any] = 4 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=__lowercase ) # check that all scores are -inf except the bos_token_id score lowerCAmelCase_ : Dict = ids_tensor((batch_size, 1) , vocab_size=2_0 ) lowerCAmelCase_ : str = 1 lowerCAmelCase_ : Optional[int] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : str = logits_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 lowerCAmelCase_ : List[Any] = 3 lowerCAmelCase_ : Union[str, Any] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : str = logits_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertFalse(jnp.isinf(__lowercase ).any() ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Tuple = 2_0 lowerCAmelCase_ : List[str] = 4 lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Any = 5 lowerCAmelCase_ : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=__lowercase , eos_token_id=__lowercase ) # check that all scores are -inf except the eos_token_id when max_length is reached lowerCAmelCase_ : List[Any] = ids_tensor((batch_size, 4) , vocab_size=2_0 ) lowerCAmelCase_ : Any = 4 lowerCAmelCase_ : List[str] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : str = logits_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached lowerCAmelCase_ : List[Any] = 3 lowerCAmelCase_ : Optional[int] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = logits_processor(__lowercase , __lowercase , cur_len=__lowercase ) self.assertFalse(jnp.isinf(__lowercase ).any() ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : str = 4 lowerCAmelCase_ : Optional[int] = 1_0 lowerCAmelCase_ : Dict = 1_5 lowerCAmelCase_ : Any = 2 lowerCAmelCase_ : str = 1 lowerCAmelCase_ : Optional[int] = 1_5 # dummy input_ids and scores lowerCAmelCase_ : Any = ids_tensor((batch_size, sequence_length) , __lowercase ) lowerCAmelCase_ : Dict = input_ids.copy() lowerCAmelCase_ : Optional[int] = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : Any = scores.copy() # instantiate all dist processors lowerCAmelCase_ : str = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowerCAmelCase_ : Union[str, Any] = FlaxTopKLogitsWarper(3 ) lowerCAmelCase_ : Any = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors lowerCAmelCase_ : List[str] = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=__lowercase ) lowerCAmelCase_ : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=__lowercase ) lowerCAmelCase_ : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=__lowercase , eos_token_id=__lowercase ) lowerCAmelCase_ : int = 1_0 # no processor list lowerCAmelCase_ : str = temp_dist_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : List[str] = top_k_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : Tuple = top_p_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : Tuple = min_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : List[str] = bos_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : str = eos_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) # with processor list lowerCAmelCase_ : Optional[int] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) lowerCAmelCase_ : Optional[Any] = processor(__lowercase , __lowercase , cur_len=__lowercase ) # scores should be equal self.assertTrue(jnp.allclose(__lowercase , __lowercase , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : List[Any] = 4 lowerCAmelCase_ : int = 1_0 lowerCAmelCase_ : int = 1_5 lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : int = 1_5 # dummy input_ids and scores lowerCAmelCase_ : Any = ids_tensor((batch_size, sequence_length) , __lowercase ) lowerCAmelCase_ : Dict = input_ids.copy() lowerCAmelCase_ : Tuple = self._get_uniform_logits(__lowercase , __lowercase ) lowerCAmelCase_ : int = scores.copy() # instantiate all dist processors lowerCAmelCase_ : Any = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowerCAmelCase_ : Any = FlaxTopKLogitsWarper(3 ) lowerCAmelCase_ : List[str] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors lowerCAmelCase_ : Dict = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=__lowercase ) lowerCAmelCase_ : List[str] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=__lowercase ) lowerCAmelCase_ : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=__lowercase , eos_token_id=__lowercase ) lowerCAmelCase_ : Dict = 1_0 # no processor list def run_no_processor_list(__lowercase , __lowercase , __lowercase ): lowerCAmelCase_ : Optional[int] = temp_dist_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : List[Any] = top_k_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : int = top_p_warp(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : int = min_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : Optional[Any] = bos_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) lowerCAmelCase_ : int = eos_dist_proc(__lowercase , __lowercase , cur_len=__lowercase ) return scores # with processor list def run_processor_list(__lowercase , __lowercase , __lowercase ): lowerCAmelCase_ : Any = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) lowerCAmelCase_ : Tuple = processor(__lowercase , __lowercase , cur_len=__lowercase ) return scores lowerCAmelCase_ : Any = jax.jit(__lowercase ) lowerCAmelCase_ : str = jax.jit(__lowercase ) lowerCAmelCase_ : Optional[int] = jitted_run_no_processor_list(__lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : str = jitted_run_processor_list(__lowercase , __lowercase , __lowercase ) # scores should be equal self.assertTrue(jnp.allclose(__lowercase , __lowercase , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
import argparse import logging import pickle from collections import Counter logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) _UpperCAmelCase : Tuple =logging.getLogger(__name__) if __name__ == "__main__": _UpperCAmelCase : str =argparse.ArgumentParser( description="""Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)""" ) parser.add_argument( """--data_file""", type=str, default="""data/dump.bert-base-uncased.pickle""", help="""The binarized dataset.""" ) parser.add_argument( """--token_counts_dump""", type=str, default="""data/token_counts.bert-base-uncased.pickle""", help="""The dump file.""" ) parser.add_argument("""--vocab_size""", default=3_0522, type=int) _UpperCAmelCase : Union[str, Any] =parser.parse_args() logger.info(f"""Loading data from {args.data_file}""") with open(args.data_file, """rb""") as fp: _UpperCAmelCase : List[str] =pickle.load(fp) logger.info("""Counting occurrences for MLM.""") _UpperCAmelCase : Any =Counter() for tk_ids in data: counter.update(tk_ids) _UpperCAmelCase : Union[str, Any] =[0] * args.vocab_size for k, v in counter.items(): _UpperCAmelCase : Dict =v logger.info(f"""Dump to {args.token_counts_dump}""") with open(args.token_counts_dump, """wb""") as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
262
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
262
1
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ = 1_000_000 )-> int: lowerCAmelCase_ : int = 0 lowerCAmelCase_ : int = 0 lowerCAmelCase_ : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f"""{solution() = }""")
262
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : str ={ """facebook/vit-mae-base""": """https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json""", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """vit_mae""" def __init__( self , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1e-12 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase=3 , __lowercase=True , __lowercase=1_6 , __lowercase=5_1_2 , __lowercase=8 , __lowercase=2_0_4_8 , __lowercase=0.75 , __lowercase=False , **__lowercase , ) -> str: super().__init__(**__lowercase ) lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : Union[str, Any] = image_size lowerCAmelCase_ : Optional[int] = patch_size lowerCAmelCase_ : Tuple = num_channels lowerCAmelCase_ : List[str] = qkv_bias lowerCAmelCase_ : List[Any] = decoder_num_attention_heads lowerCAmelCase_ : int = decoder_hidden_size lowerCAmelCase_ : Optional[int] = decoder_num_hidden_layers lowerCAmelCase_ : Tuple = decoder_intermediate_size lowerCAmelCase_ : Tuple = mask_ratio lowerCAmelCase_ : Any = norm_pix_loss
262
1
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCAmelCase : Union[str, Any] =logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[str] = field( default="""cifar10""", metadata={"""help""": """Name of a dataset from the datasets package"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """The column name of the images in the files."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field(default=UpperCAmelCase__, metadata={"""help""": """A folder containing the training data."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field(default=UpperCAmelCase__, metadata={"""help""": """A folder containing the validation data."""} ) SCREAMING_SNAKE_CASE__ : Optional[float] = field( default=0.15, metadata={"""help""": """Percent to split off of train for validation."""} ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Union[str, Any] = {} if self.train_dir is not None: lowerCAmelCase_ : int = self.train_dir if self.validation_dir is not None: lowerCAmelCase_ : int = self.validation_dir lowerCAmelCase_ : Optional[Any] = data_files if data_files else None @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = field( default=UpperCAmelCase__, metadata={ """help""": ( """The model checkpoint for weights initialization.Don't set if you want to train a model from scratch.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained config name or path if not the same as model_name_or_path"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Where do you want to store the pretrained models downloaded from s3"""} ) SCREAMING_SNAKE_CASE__ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) SCREAMING_SNAKE_CASE__ : str = field(default=UpperCAmelCase__, metadata={"""help""": """Name or path of preprocessor config."""} ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) SCREAMING_SNAKE_CASE__ : float = field( default=0.75, metadata={"""help""": """The ratio of the number of masked tokens in the input sequence."""} ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={"""help""": """Whether or not to train with normalized pixel values as target."""} ) @dataclass class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : float = field( default=1e-3, metadata={"""help""": """Base learning rate: absolute_lr = base_lr * total_batch_size / 256."""} ) def lowerCAmelCase ( lowerCAmelCase_ )-> Dict: lowerCAmelCase_ : Union[str, Any] = torch.stack([example['''pixel_values'''] for example in examples] ) return {"pixel_values": pixel_values} def lowerCAmelCase ( )-> Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_mae''' , lowerCAmelCase_ , lowerCAmelCase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCAmelCase_ : str = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase_ ) transformers.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCAmelCase_ : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase_ : Tuple = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset. lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. lowerCAmelCase_ : Optional[int] = None if '''validation''' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowerCAmelCase_ ) and data_args.train_val_split > 0.0: lowerCAmelCase_ : List[str] = ds['''train'''].train_test_split(data_args.train_val_split ) lowerCAmelCase_ : Optional[int] = split['''train'''] lowerCAmelCase_ : List[str] = split['''test'''] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase_ : int = { '''cache_dir''': model_args.cache_dir, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.config_name: lowerCAmelCase_ : Optional[int] = ViTMAEConfig.from_pretrained(model_args.config_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: lowerCAmelCase_ : Union[str, Any] = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: lowerCAmelCase_ : Any = ViTMAEConfig() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(f"""Overriding config: {model_args.config_overrides}""" ) config.update_from_string(model_args.config_overrides ) logger.info(f"""New config: {config}""" ) # adapt config config.update( { '''mask_ratio''': model_args.mask_ratio, '''norm_pix_loss''': model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: lowerCAmelCase_ : Union[str, Any] = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: lowerCAmelCase_ : int = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: lowerCAmelCase_ : int = ViTImageProcessor() # create model if model_args.model_name_or_path: lowerCAmelCase_ : str = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('''Training new model from scratch''' ) lowerCAmelCase_ : str = ViTMAEForPreTraining(lowerCAmelCase_ ) if training_args.do_train: lowerCAmelCase_ : Union[str, Any] = ds['''train'''].column_names else: lowerCAmelCase_ : int = ds['''validation'''].column_names if data_args.image_column_name is not None: lowerCAmelCase_ : Dict = data_args.image_column_name elif "image" in column_names: lowerCAmelCase_ : List[str] = '''image''' elif "img" in column_names: lowerCAmelCase_ : Union[str, Any] = '''img''' else: lowerCAmelCase_ : List[str] = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: lowerCAmelCase_ : Optional[int] = image_processor.size['''shortest_edge'''] else: lowerCAmelCase_ : str = (image_processor.size['''height'''], image_processor.size['''width''']) lowerCAmelCase_ : List[str] = Compose( [ Lambda(lambda lowerCAmelCase_ : img.convert('''RGB''' ) if img.mode != "RGB" else img ), RandomResizedCrop(lowerCAmelCase_ , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(lowerCAmelCase_ ): lowerCAmelCase_ : Union[str, Any] = [transforms(lowerCAmelCase_ ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError('''--do_train requires a train dataset''' ) if data_args.max_train_samples is not None: lowerCAmelCase_ : List[str] = ds['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(lowerCAmelCase_ ) if training_args.do_eval: if "validation" not in ds: raise ValueError('''--do_eval requires a validation dataset''' ) if data_args.max_eval_samples is not None: lowerCAmelCase_ : Optional[Any] = ( ds['''validation'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(lowerCAmelCase_ ) # Compute absolute learning rate lowerCAmelCase_ : Tuple = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: lowerCAmelCase_ : Tuple = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer lowerCAmelCase_ : Optional[int] = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=ds['''train'''] if training_args.do_train else None , eval_dataset=ds['''validation'''] if training_args.do_eval else None , tokenizer=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , ) # Training if training_args.do_train: lowerCAmelCase_ : str = None if training_args.resume_from_checkpoint is not None: lowerCAmelCase_ : Any = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCAmelCase_ : Any = last_checkpoint lowerCAmelCase_ : List[Any] = trainer.train(resume_from_checkpoint=lowerCAmelCase_ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: lowerCAmelCase_ : Tuple = trainer.evaluate() trainer.log_metrics('''eval''' , lowerCAmelCase_ ) trainer.save_metrics('''eval''' , lowerCAmelCase_ ) # Write model card and (optionally) push to hub lowerCAmelCase_ : Optional[int] = { '''tasks''': '''masked-auto-encoding''', '''dataset''': data_args.dataset_name, '''tags''': ['''masked-auto-encoding'''], } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase_ ) else: trainer.create_model_card(**lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> str: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
def lowerCAmelCase ( lowerCAmelCase_ = 10**9 )-> int: lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : str = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value lowerCAmelCase_ : Any = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f"""{solution() = }""")
262
1
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _UpperCAmelCase : Dict =argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument( """--original_config_file""", type=str, required=True, help="""The YAML config file corresponding to the original architecture.""", ) parser.add_argument( """--num_in_channels""", default=None, type=int, help="""The number of input channels. If `None` number of input channels will be automatically inferred.""", ) parser.add_argument( """--image_size""", default=512, type=int, help=( """The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2""" """ Base. Use 768 for Stable Diffusion v2.""" ), ) parser.add_argument( """--extract_ema""", action="""store_true""", help=( """Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights""" """ or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield""" """ higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.""" ), ) parser.add_argument( """--upcast_attention""", action="""store_true""", help=( """Whether the attention computation should always be upcasted. This is necessary when running stable""" """ diffusion 2.1.""" ), ) parser.add_argument( """--from_safetensors""", action="""store_true""", help="""If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.""", ) parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") def lowerCAmelCase ( lowerCAmelCase_ )-> str: if string == "True": return True elif string == "False": return False else: raise ValueError(f"""could not parse string as bool {string}""" ) parser.add_argument( """--use_linear_projection""", help="""Override for use linear projection""", required=False, type=parse_bool ) parser.add_argument("""--cross_attention_dim""", help="""Override for cross attention_dim""", required=False, type=int) _UpperCAmelCase : Optional[Any] =parser.parse_args() _UpperCAmelCase : List[Any] =download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
262
import inspect import unittest class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: try: import diffusers # noqa: F401 except ImportError: assert False def lowercase_ ( self ) -> List[str]: import diffusers from diffusers.dependency_versions_table import deps lowerCAmelCase_ : Any = inspect.getmembers(__lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowerCAmelCase_ : Optional[int] = '''k-diffusion''' elif backend == "invisible_watermark": lowerCAmelCase_ : Dict = '''invisible-watermark''' assert backend in deps, f"""{backend} is not in the deps table!"""
262
1
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCAmelCase : Any =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
1
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _UpperCAmelCase : Any ={ """iou_prediction_head.layers.0""": """iou_prediction_head.proj_in""", """iou_prediction_head.layers.1""": """iou_prediction_head.layers.0""", """iou_prediction_head.layers.2""": """iou_prediction_head.proj_out""", """mask_decoder.output_upscaling.0""": """mask_decoder.upscale_conv1""", """mask_decoder.output_upscaling.1""": """mask_decoder.upscale_layer_norm""", """mask_decoder.output_upscaling.3""": """mask_decoder.upscale_conv2""", """mask_downscaling.0""": """mask_embed.conv1""", """mask_downscaling.1""": """mask_embed.layer_norm1""", """mask_downscaling.3""": """mask_embed.conv2""", """mask_downscaling.4""": """mask_embed.layer_norm2""", """mask_downscaling.6""": """mask_embed.conv3""", """point_embeddings""": """point_embed""", """pe_layer.positional_encoding_gaussian_matrix""": """shared_embedding.positional_embedding""", """image_encoder""": """vision_encoder""", """neck.0""": """neck.conv1""", """neck.1""": """neck.layer_norm1""", """neck.2""": """neck.conv2""", """neck.3""": """neck.layer_norm2""", """patch_embed.proj""": """patch_embed.projection""", """.norm""": """.layer_norm""", """blocks""": """layers""", } def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: lowerCAmelCase_ : Optional[int] = {} state_dict.pop('''pixel_mean''' , lowerCAmelCase_ ) state_dict.pop('''pixel_std''' , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = r'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : List[str] = key.replace(lowerCAmelCase_ , lowerCAmelCase_ ) if re.match(lowerCAmelCase_ , lowerCAmelCase_ ): lowerCAmelCase_ : Tuple = int(re.match(lowerCAmelCase_ , lowerCAmelCase_ ).group(2 ) ) if layer_nb == 0: lowerCAmelCase_ : Optional[Any] = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: lowerCAmelCase_ : Optional[Any] = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: lowerCAmelCase_ : int = key.replace('''layers.2''' , '''proj_out''' ) lowerCAmelCase_ : List[Any] = value lowerCAmelCase_ : Optional[Any] = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="ybelkada/segment-anything" )-> Optional[Any]: lowerCAmelCase_ : Dict = hf_hub_download(lowerCAmelCase_ , f"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: lowerCAmelCase_ : Dict = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : int = SamVisionConfig( hidden_size=1_024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=lowerCAmelCase_ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : List[str] = SamVisionConfig( hidden_size=1_280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=lowerCAmelCase_ , ) lowerCAmelCase_ : Union[str, Any] = torch.load(lowerCAmelCase_ , map_location='''cpu''' ) lowerCAmelCase_ : Tuple = replace_keys(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = SamImageProcessor() lowerCAmelCase_ : Optional[int] = SamProcessor(image_processor=lowerCAmelCase_ ) lowerCAmelCase_ : int = SamModel(lowerCAmelCase_ ) hf_model.load_state_dict(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = hf_model.to('''cuda''' ) lowerCAmelCase_ : List[str] = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' lowerCAmelCase_ : Tuple = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ).convert('''RGB''' ) lowerCAmelCase_ : Optional[Any] = [[[400, 650]]] lowerCAmelCase_ : Optional[int] = [[1]] lowerCAmelCase_ : Any = processor(images=np.array(lowerCAmelCase_ ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**lowerCAmelCase_ ) lowerCAmelCase_ : str = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 lowerCAmelCase_ : Union[str, Any] = processor( images=np.array(lowerCAmelCase_ ) , input_points=lowerCAmelCase_ , input_labels=lowerCAmelCase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase_ : Optional[int] = hf_model(**lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 lowerCAmelCase_ : List[str] = ((75, 275, 1_725, 850),) lowerCAmelCase_ : Dict = processor(images=np.array(lowerCAmelCase_ ) , input_boxes=lowerCAmelCase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase_ : List[str] = hf_model(**lowerCAmelCase_ ) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. lowerCAmelCase_ : Union[str, Any] = [[[400, 650], [800, 650]]] lowerCAmelCase_ : List[Any] = [[1, 1]] lowerCAmelCase_ : int = processor( images=np.array(lowerCAmelCase_ ) , input_points=lowerCAmelCase_ , input_labels=lowerCAmelCase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": _UpperCAmelCase : Optional[Any] =argparse.ArgumentParser() _UpperCAmelCase : List[Any] =["""sam_vit_b_01ec64""", """sam_vit_h_4b8939""", """sam_vit_l_0b3195"""] parser.add_argument( """--model_name""", default="""sam_vit_h_4b8939""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) parser.add_argument( """--model_hub_id""", default="""ybelkada/segment-anything""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) _UpperCAmelCase : int =parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
262
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
1
import unittest from knapsack import greedy_knapsack as kp class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: lowerCAmelCase_ : Dict = [1_0, 2_0, 3_0, 4_0, 5_0, 6_0] lowerCAmelCase_ : Tuple = [2, 4, 6, 8, 1_0, 1_2] lowerCAmelCase_ : Optional[int] = 1_0_0 self.assertEqual(kp.calc_profit(__lowercase , __lowercase , __lowercase ) , 2_1_0 ) def lowercase_ ( self ) -> List[Any]: self.assertRaisesRegex(__lowercase , '''max_weight must greater than zero.''' ) def lowercase_ ( self ) -> Tuple: self.assertRaisesRegex(__lowercase , '''Weight can not be negative.''' ) def lowercase_ ( self ) -> int: self.assertRaisesRegex(__lowercase , '''Profit can not be negative.''' ) def lowercase_ ( self ) -> Dict: self.assertRaisesRegex(__lowercase , '''max_weight must greater than zero.''' ) def lowercase_ ( self ) -> Optional[int]: self.assertRaisesRegex( __lowercase , '''The length of profit and weight must be same.''' ) if __name__ == "__main__": unittest.main()
262
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : str =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ={ """sayakpaul/vit-msn-base""": """https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json""", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = """vit_msn""" def __init__( self , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1e-06 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase=3 , __lowercase=True , **__lowercase , ) -> List[str]: super().__init__(**__lowercase ) lowerCAmelCase_ : List[str] = hidden_size lowerCAmelCase_ : Optional[Any] = num_hidden_layers lowerCAmelCase_ : List[str] = num_attention_heads lowerCAmelCase_ : Tuple = intermediate_size lowerCAmelCase_ : Optional[Any] = hidden_act lowerCAmelCase_ : List[Any] = hidden_dropout_prob lowerCAmelCase_ : Dict = attention_probs_dropout_prob lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Optional[int] = layer_norm_eps lowerCAmelCase_ : List[Any] = image_size lowerCAmelCase_ : Any = patch_size lowerCAmelCase_ : List[Any] = num_channels lowerCAmelCase_ : Optional[int] = qkv_bias
262
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
1
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> np.ndarray: # prepare kernel # the kernel size have to be odd if (ksize % 2) == 0: lowerCAmelCase_ : Any = ksize + 1 lowerCAmelCase_ : Any = np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(lowerCAmelCase_ ): for x in range(lowerCAmelCase_ ): # distance from center lowerCAmelCase_ : Any = x - ksize // 2 lowerCAmelCase_ : str = y - ksize // 2 # degree to radiant lowerCAmelCase_ : Optional[int] = theta / 180 * np.pi lowerCAmelCase_ : str = np.cos(_theta ) lowerCAmelCase_ : int = np.sin(_theta ) # get kernel x lowerCAmelCase_ : Dict = cos_theta * px + sin_theta * py # get kernel y lowerCAmelCase_ : Any = -sin_theta * px + cos_theta * py # fill kernel lowerCAmelCase_ : List[str] = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image _UpperCAmelCase : Optional[int] =imread("""../image_data/lena.jpg""") # turn image in gray scale value _UpperCAmelCase : List[Any] =cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges _UpperCAmelCase : List[str] =np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 120, 150]: _UpperCAmelCase : Tuple =gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) _UpperCAmelCase : int =out / out.max() * 255 _UpperCAmelCase : str =out.astype(np.uinta) imshow("""Original""", gray) imshow("""Gabor filter with 20x20 mask and 6 directions""", out) waitKey(0)
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, get_image_size, infer_channel_dimension_format, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_vision_available, logging from ...utils.import_utils import requires_backends if is_vision_available(): import textwrap from PIL import Image, ImageDraw, ImageFont if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: _UpperCAmelCase : Union[str, Any] =False _UpperCAmelCase : Any =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ="""ybelkada/fonts""" def lowerCAmelCase ( )-> Dict: if is_torch_available() and not is_torch_greater_or_equal_than_1_11: raise ImportError( f"""You are using torch=={torch.__version__}, but torch>=1.11.0 is required to use """ '''Pix2StructImageProcessor. Please upgrade torch.''' ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: requires_backends(lowerCAmelCase_ , ['''torch'''] ) _check_torch_version() lowerCAmelCase_ : int = image_tensor.unsqueeze(0 ) lowerCAmelCase_ : Optional[int] = torch.nn.functional.unfold(lowerCAmelCase_ , (patch_height, patch_width) , stride=(patch_height, patch_width) ) lowerCAmelCase_ : List[str] = patches.reshape(image_tensor.size(0 ) , image_tensor.size(1 ) , lowerCAmelCase_ , lowerCAmelCase_ , -1 ) lowerCAmelCase_ : List[str] = patches.permute(0 , 4 , 2 , 3 , 1 ).reshape( image_tensor.size(2 ) // patch_height , image_tensor.size(3 ) // patch_width , image_tensor.size(1 ) * patch_height * patch_width , ) return patches.unsqueeze(0 ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = 36 , lowerCAmelCase_ = "black" , lowerCAmelCase_ = "white" , lowerCAmelCase_ = 5 , lowerCAmelCase_ = 5 , lowerCAmelCase_ = 5 , lowerCAmelCase_ = 5 , lowerCAmelCase_ = None , lowerCAmelCase_ = None , )-> Image.Image: requires_backends(lowerCAmelCase_ , '''vision''' ) # Add new lines so that each line is no more than 80 characters. lowerCAmelCase_ : Optional[int] = textwrap.TextWrapper(width=80 ) lowerCAmelCase_ : int = wrapper.wrap(text=lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = '''\n'''.join(lowerCAmelCase_ ) if font_bytes is not None and font_path is None: lowerCAmelCase_ : str = io.BytesIO(lowerCAmelCase_ ) elif font_path is not None: lowerCAmelCase_ : Union[str, Any] = font_path else: lowerCAmelCase_ : List[Any] = hf_hub_download(lowerCAmelCase_ , '''Arial.TTF''' ) lowerCAmelCase_ : Any = ImageFont.truetype(lowerCAmelCase_ , encoding='''UTF-8''' , size=lowerCAmelCase_ ) # Use a temporary canvas to determine the width and height in pixels when # rendering the text. lowerCAmelCase_ : Tuple = ImageDraw.Draw(Image.new('''RGB''' , (1, 1) , lowerCAmelCase_ ) ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = temp_draw.textbbox((0, 0) , lowerCAmelCase_ , lowerCAmelCase_ ) # Create the actual image with a bit of padding around the text. lowerCAmelCase_ : List[str] = text_width + left_padding + right_padding lowerCAmelCase_ : int = text_height + top_padding + bottom_padding lowerCAmelCase_ : List[str] = Image.new('''RGB''' , (image_width, image_height) , lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = ImageDraw.Draw(lowerCAmelCase_ ) draw.text(xy=(left_padding, top_padding) , text=lowerCAmelCase_ , fill=lowerCAmelCase_ , font=lowerCAmelCase_ ) return image def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ )-> Optional[Any]: requires_backends(lowerCAmelCase_ , '''vision''' ) # Convert to PIL image if necessary lowerCAmelCase_ : Tuple = to_pil_image(lowerCAmelCase_ ) lowerCAmelCase_ : List[Any] = render_text(lowerCAmelCase_ , **lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(header_image.width , image.width ) lowerCAmelCase_ : List[str] = int(image.height * (new_width / image.width) ) lowerCAmelCase_ : Tuple = int(header_image.height * (new_width / header_image.width) ) lowerCAmelCase_ : Optional[int] = Image.new('''RGB''' , (new_width, new_height + new_header_height) , '''white''' ) new_image.paste(header_image.resize((new_width, new_header_height) ) , (0, 0) ) new_image.paste(image.resize((new_width, new_height) ) , (0, new_header_height) ) # Convert back to the original framework if necessary lowerCAmelCase_ : Dict = to_numpy_array(lowerCAmelCase_ ) if infer_channel_dimension_format(lowerCAmelCase_ ) == ChannelDimension.LAST: lowerCAmelCase_ : int = to_channel_dimension_format(lowerCAmelCase_ , ChannelDimension.LAST ) return new_image class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = ["""flattened_patches"""] def __init__( self , __lowercase = True , __lowercase = True , __lowercase = None , __lowercase = 2_0_4_8 , __lowercase = False , **__lowercase , ) -> None: super().__init__(**__lowercase ) lowerCAmelCase_ : List[Any] = patch_size if patch_size is not None else {'''height''': 1_6, '''width''': 1_6} lowerCAmelCase_ : List[Any] = do_normalize lowerCAmelCase_ : List[Any] = do_convert_rgb lowerCAmelCase_ : Tuple = max_patches lowerCAmelCase_ : Optional[Any] = is_vqa def lowercase_ ( self , __lowercase , __lowercase , __lowercase , **__lowercase ) -> np.ndarray: requires_backends(self.extract_flattened_patches , '''torch''' ) _check_torch_version() # convert to torch lowerCAmelCase_ : int = to_channel_dimension_format(__lowercase , ChannelDimension.FIRST ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(__lowercase ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = patch_size['''height'''], patch_size['''width'''] lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = get_image_size(__lowercase ) # maximize scale s.t. lowerCAmelCase_ : Union[str, Any] = math.sqrt(max_patches * (patch_height / image_height) * (patch_width / image_width) ) lowerCAmelCase_ : Optional[int] = max(min(math.floor(scale * image_height / patch_height ) , __lowercase ) , 1 ) lowerCAmelCase_ : int = max(min(math.floor(scale * image_width / patch_width ) , __lowercase ) , 1 ) lowerCAmelCase_ : Union[str, Any] = max(num_feasible_rows * patch_height , 1 ) lowerCAmelCase_ : Union[str, Any] = max(num_feasible_cols * patch_width , 1 ) lowerCAmelCase_ : List[str] = torch.nn.functional.interpolate( image.unsqueeze(0 ) , size=(resized_height, resized_width) , mode='''bilinear''' , align_corners=__lowercase , antialias=__lowercase , ).squeeze(0 ) # [1, rows, columns, patch_height * patch_width * image_channels] lowerCAmelCase_ : Any = torch_extract_patches(__lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Dict = patches.shape lowerCAmelCase_ : Optional[int] = patches_shape[1] lowerCAmelCase_ : str = patches_shape[2] lowerCAmelCase_ : List[Any] = patches_shape[3] # [rows * columns, patch_height * patch_width * image_channels] lowerCAmelCase_ : Optional[int] = patches.reshape([rows * columns, depth] ) # [rows * columns, 1] lowerCAmelCase_ : Any = torch.arange(__lowercase ).reshape([rows, 1] ).repeat(1 , __lowercase ).reshape([rows * columns, 1] ) lowerCAmelCase_ : Optional[int] = torch.arange(__lowercase ).reshape([1, columns] ).repeat(__lowercase , 1 ).reshape([rows * columns, 1] ) # Offset by 1 so the ids do not contain zeros, which represent padding. row_ids += 1 col_ids += 1 # Prepare additional patch features. # [rows * columns, 1] lowerCAmelCase_ : List[Any] = row_ids.to(torch.floataa ) lowerCAmelCase_ : Any = col_ids.to(torch.floataa ) # [rows * columns, 2 + patch_height * patch_width * image_channels] lowerCAmelCase_ : List[str] = torch.cat([row_ids, col_ids, patches] , -1 ) # [max_patches, 2 + patch_height * patch_width * image_channels] lowerCAmelCase_ : int = torch.nn.functional.pad(__lowercase , [0, 0, 0, max_patches - (rows * columns)] ).float() lowerCAmelCase_ : Optional[int] = to_numpy_array(__lowercase ) return result def lowercase_ ( self , __lowercase , __lowercase = None , **__lowercase ) -> np.ndarray: if image.dtype == np.uinta: lowerCAmelCase_ : Tuple = image.astype(np.floataa ) # take mean across the whole `image` lowerCAmelCase_ : Dict = np.mean(__lowercase ) lowerCAmelCase_ : Optional[Any] = np.std(__lowercase ) lowerCAmelCase_ : Union[str, Any] = max(__lowercase , 1.0 / math.sqrt(np.prod(image.shape ) ) ) return normalize(__lowercase , mean=__lowercase , std=__lowercase , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> ImageInput: lowerCAmelCase_ : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase_ : Union[str, Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCAmelCase_ : Any = patch_size if patch_size is not None else self.patch_size lowerCAmelCase_ : Union[str, Any] = max_patches if max_patches is not None else self.max_patches lowerCAmelCase_ : Optional[int] = self.is_vqa if kwargs.get('''data_format''' , __lowercase ) is not None: raise ValueError('''data_format is not an accepted input as the outputs are ''' ) lowerCAmelCase_ : str = make_list_of_images(__lowercase ) if not valid_images(__lowercase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCAmelCase_ : Any = [convert_to_rgb(__lowercase ) for image in images] # All transformations expect numpy arrays. lowerCAmelCase_ : Tuple = [to_numpy_array(__lowercase ) for image in images] if is_vqa: if header_text is None: raise ValueError('''A header text must be provided for VQA models.''' ) lowerCAmelCase_ : Union[str, Any] = kwargs.pop('''font_bytes''' , __lowercase ) lowerCAmelCase_ : Union[str, Any] = kwargs.pop('''font_path''' , __lowercase ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Any = [header_text] * len(__lowercase ) lowerCAmelCase_ : int = [ render_header(__lowercase , header_text[i] , font_bytes=__lowercase , font_path=__lowercase ) for i, image in enumerate(__lowercase ) ] if do_normalize: lowerCAmelCase_ : str = [self.normalize(image=__lowercase ) for image in images] # convert to torch tensor and permute lowerCAmelCase_ : List[Any] = [ self.extract_flattened_patches(image=__lowercase , max_patches=__lowercase , patch_size=__lowercase ) for image in images ] # create attention mask in numpy lowerCAmelCase_ : str = [(image.sum(axis=-1 ) != 0).astype(np.floataa ) for image in images] lowerCAmelCase_ : List[Any] = BatchFeature( data={'''flattened_patches''': images, '''attention_mask''': attention_masks} , tensor_type=__lowercase ) return encoded_outputs
262
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
1
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed _UpperCAmelCase : List[str] ={ """distilbert""": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), """roberta""": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), """bert""": (BertConfig, BertForMaskedLM, BertTokenizer), """gpt2""": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def lowerCAmelCase ( lowerCAmelCase_ )-> Any: assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[Any]: if args.student_type == "roberta": lowerCAmelCase_ : Tuple = False elif args.student_type == "gpt2": lowerCAmelCase_ : List[str] = False def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: if args.student_type == "roberta": lowerCAmelCase_ : List[str] = False def lowerCAmelCase ( )-> Tuple: lowerCAmelCase_ : Dict = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=lowerCAmelCase_ , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=lowerCAmelCase_ , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=lowerCAmelCase_ , type=lowerCAmelCase_ , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=lowerCAmelCase_ , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=lowerCAmelCase_ , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=lowerCAmelCase_ , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=lowerCAmelCase_ , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=lowerCAmelCase_ , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=lowerCAmelCase_ , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=lowerCAmelCase_ , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.15 , type=lowerCAmelCase_ , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=lowerCAmelCase_ , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=lowerCAmelCase_ , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=lowerCAmelCase_ , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=lowerCAmelCase_ , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=lowerCAmelCase_ , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=lowerCAmelCase_ , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=lowerCAmelCase_ , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=lowerCAmelCase_ , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.05 , type=lowerCAmelCase_ , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=lowerCAmelCase_ , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5e-4 , type=lowerCAmelCase_ , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1e-6 , type=lowerCAmelCase_ , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=lowerCAmelCase_ , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.02 , type=lowerCAmelCase_ , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=lowerCAmelCase_ , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=lowerCAmelCase_ , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=lowerCAmelCase_ , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=lowerCAmelCase_ , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=lowerCAmelCase_ , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=lowerCAmelCase_ , default=4_000 , help='''Checkpoint interval.''' ) lowerCAmelCase_ : Dict = parser.parse_args() sanity_checks(lowerCAmelCase_ ) # ARGS # init_gpu_params(lowerCAmelCase_ ) set_seed(lowerCAmelCase_ ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(f"""Param: {args}""" ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(lowerCAmelCase_ ) , lowerCAmelCase_ , indent=4 ) git_log(args.dump_path ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = MODEL_CLASSES[args.student_type] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = MODEL_CLASSES[args.teacher_type] # TOKENIZER # lowerCAmelCase_ : Any = teacher_tokenizer_class.from_pretrained(args.teacher_name ) lowerCAmelCase_ : Dict = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): lowerCAmelCase_ : Tuple = tokenizer.all_special_tokens.index(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = tokenizer.all_special_ids[idx] logger.info(f"""Special tokens {special_tok_ids}""" ) lowerCAmelCase_ : List[Any] = special_tok_ids lowerCAmelCase_ : Tuple = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"""Loading data from {args.data_file}""" ) with open(args.data_file , '''rb''' ) as fp: lowerCAmelCase_ : Optional[Any] = pickle.load(lowerCAmelCase_ ) if args.mlm: logger.info(f"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts , '''rb''' ) as fp: lowerCAmelCase_ : List[str] = pickle.load(lowerCAmelCase_ ) lowerCAmelCase_ : int = np.maximum(lowerCAmelCase_ , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): lowerCAmelCase_ : Any = 0.0 # do not predict special tokens lowerCAmelCase_ : List[Any] = torch.from_numpy(lowerCAmelCase_ ) else: lowerCAmelCase_ : Optional[int] = None lowerCAmelCase_ : Union[str, Any] = LmSeqsDataset(params=lowerCAmelCase_ , data=lowerCAmelCase_ ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f"""Loading student config from {args.student_config}""" ) lowerCAmelCase_ : List[Any] = student_config_class.from_pretrained(args.student_config ) lowerCAmelCase_ : str = True if args.student_pretrained_weights is not None: logger.info(f"""Loading pretrained weights from {args.student_pretrained_weights}""" ) lowerCAmelCase_ : Dict = student_model_class.from_pretrained(args.student_pretrained_weights , config=lowerCAmelCase_ ) else: lowerCAmelCase_ : List[Any] = student_model_class(lowerCAmelCase_ ) if args.n_gpu > 0: student.to(f"""cuda:{args.local_rank}""" ) logger.info('''Student loaded.''' ) # TEACHER # lowerCAmelCase_ : List[str] = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=lowerCAmelCase_ ) if args.n_gpu > 0: teacher.to(f"""cuda:{args.local_rank}""" ) logger.info(f"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(lowerCAmelCase_ , lowerCAmelCase_ ) if args.freeze_token_type_embds: freeze_token_type_embeddings(lowerCAmelCase_ , lowerCAmelCase_ ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() lowerCAmelCase_ : int = Distiller( params=lowerCAmelCase_ , dataset=lowerCAmelCase_ , token_probs=lowerCAmelCase_ , student=lowerCAmelCase_ , teacher=lowerCAmelCase_ ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
262
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def lowerCAmelCase ( )-> None: print('''Making key files...''' ) make_key_files('''rsa''' , 1_024 ) print('''Key files generation successful.''' ) def lowerCAmelCase ( lowerCAmelCase_ )-> tuple[tuple[int, int], tuple[int, int]]: print('''Generating prime p...''' ) lowerCAmelCase_ : Optional[Any] = rabinMiller.generate_large_prime(lowerCAmelCase_ ) print('''Generating prime q...''' ) lowerCAmelCase_ : str = rabinMiller.generate_large_prime(lowerCAmelCase_ ) lowerCAmelCase_ : int = p * q print('''Generating e that is relatively prime to (p - 1) * (q - 1)...''' ) while True: lowerCAmelCase_ : str = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(lowerCAmelCase_ , (p - 1) * (q - 1) ) == 1: break print('''Calculating d that is mod inverse of e...''' ) lowerCAmelCase_ : Optional[int] = cryptoMath.find_mod_inverse(lowerCAmelCase_ , (p - 1) * (q - 1) ) lowerCAmelCase_ : Tuple = (n, e) lowerCAmelCase_ : List[str] = (n, d) return (public_key, private_key) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> None: if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('''\nWARNING:''' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" '''Use a different name or delete these files and re-run this program.''' ) sys.exit() lowerCAmelCase_ , lowerCAmelCase_ : int = generate_key(lowerCAmelCase_ ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , '''w''' ) as out_file: out_file.write(f"""{key_size},{public_key[0]},{public_key[1]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , '''w''' ) as out_file: out_file.write(f"""{key_size},{private_key[0]},{private_key[1]}""" ) if __name__ == "__main__": main()
262
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
1
def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> float: def get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) -> str: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : Union[str, Any] = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): lowerCAmelCase_ : str = int(max(0 , i - limit ) ) lowerCAmelCase_ : List[Any] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = f"""{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}""" return "".join(lowerCAmelCase_ ) # matching characters lowerCAmelCase_ : int = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : str = len(lowerCAmelCase_ ) # transposition lowerCAmelCase_ : List[Any] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: lowerCAmelCase_ : List[Any] = 0.0 else: lowerCAmelCase_ : str = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters lowerCAmelCase_ : int = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("""hello""", """world"""))
262
from __future__ import annotations import math class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : str = size # approximate the overall size of segment tree with given value lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] lowerCAmelCase_ : Optional[int] = [0 for i in range(0 , 4 * size )] # flag for lazy update def lowercase_ ( self , __lowercase ) -> int: return idx * 2 def lowercase_ ( self , __lowercase ) -> int: return idx * 2 + 1 def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> None: if left_element == right_element: lowerCAmelCase_ : Tuple = a[left_element - 1] else: lowerCAmelCase_ : int = (left_element + right_element) // 2 self.build(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase ) self.build(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase ) lowerCAmelCase_ : Any = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> bool: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Union[str, Any] = False if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Any = self.lazy[idx] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCAmelCase_ : Dict = val if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = val lowerCAmelCase_ : List[Any] = val lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : List[str] = True return True lowerCAmelCase_ : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) self.update(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) return True def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> int | float: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Optional[Any] = False if left_element != right_element: lowerCAmelCase_ : List[Any] = self.lazy[idx] lowerCAmelCase_ : Dict = self.lazy[idx] lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCAmelCase_ : List[Any] = (left_element + right_element) // 2 lowerCAmelCase_ : Tuple = self.query(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.query(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase ) return max(__lowercase , __lowercase ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , __lowercase , __lowercase ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _UpperCAmelCase : str =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _UpperCAmelCase : List[str] =15 _UpperCAmelCase : Any =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
262
1
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar _UpperCAmelCase : Tuple =TypeVar("""T""") class snake_case__( Generic[T] ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : deque[T] # Cache store of keys SCREAMING_SNAKE_CASE__ : set[T] # References of the keys in cache SCREAMING_SNAKE_CASE__ : int = 10 # Maximum capacity of cache def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : Optional[int] = deque() lowerCAmelCase_ : List[str] = set() if not n: lowerCAmelCase_ : List[Any] = sys.maxsize elif n < 0: raise ValueError('''n should be an integer greater than 0.''' ) else: lowerCAmelCase_ : Dict = n def lowercase_ ( self , __lowercase ) -> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: lowerCAmelCase_ : List[Any] = self.dq_store.pop() self.key_reference.remove(__lowercase ) else: self.dq_store.remove(__lowercase ) self.dq_store.appendleft(__lowercase ) self.key_reference.add(__lowercase ) def lowercase_ ( self ) -> None: for k in self.dq_store: print(__lowercase ) def __repr__( self ) -> str: return f"""LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}""" if __name__ == "__main__": import doctest doctest.testmod() _UpperCAmelCase : LRUCache[str | int] =LRUCache(4) lru_cache.refer("""A""") lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer("""A""") lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
262
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py _UpperCAmelCase : Optional[int] ="""src/transformers""" _UpperCAmelCase : str ="""docs/source/en""" _UpperCAmelCase : Optional[int] =""".""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : int = f.readlines() # Find the start prompt. lowerCAmelCase_ : List[Any] = 0 while not lines[start_index].startswith(lowerCAmelCase_ ): start_index += 1 start_index += 1 lowerCAmelCase_ : List[str] = start_index while not lines[end_index].startswith(lowerCAmelCase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | _UpperCAmelCase : Optional[Any] ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. _UpperCAmelCase : Optional[int] =re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _UpperCAmelCase : Dict =re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _UpperCAmelCase : Optional[Any] =re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : Optional[int] =direct_transformers_import(TRANSFORMERS_PATH) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : str = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' , lowerCAmelCase_ ) return [m.group(0 ) for m in matches] def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Tuple = 2 if text == '''✅''' or text == '''❌''' else len(lowerCAmelCase_ ) lowerCAmelCase_ : int = (width - text_length) // 2 lowerCAmelCase_ : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def lowerCAmelCase ( )-> str: lowerCAmelCase_ : Any = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowerCAmelCase_ : Dict = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } lowerCAmelCase_ : List[Any] = {name: config.replace('''Config''' , '''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. lowerCAmelCase_ : Tuple = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = None if attr_name.endswith('''Tokenizer''' ): lowerCAmelCase_ : Union[str, Any] = slow_tokenizers lowerCAmelCase_ : List[str] = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): lowerCAmelCase_ : int = fast_tokenizers lowerCAmelCase_ : Union[str, Any] = attr_name[:-13] elif _re_tf_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = tf_models lowerCAmelCase_ : str = _re_tf_models.match(lowerCAmelCase_ ).groups()[0] elif _re_flax_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = flax_models lowerCAmelCase_ : Union[str, Any] = _re_flax_models.match(lowerCAmelCase_ ).groups()[0] elif _re_pt_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Any = pt_models lowerCAmelCase_ : List[Any] = _re_pt_models.match(lowerCAmelCase_ ).groups()[0] if lookup_dict is not None: while len(lowerCAmelCase_ ) > 0: if attr_name in model_name_to_prefix.values(): lowerCAmelCase_ : Union[str, Any] = True break # Try again after removing the last word in the name lowerCAmelCase_ : Any = ''''''.join(camel_case_split(lowerCAmelCase_ )[:-1] ) # Let's build that table! lowerCAmelCase_ : int = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) lowerCAmelCase_ : Tuple = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). lowerCAmelCase_ : Union[str, Any] = [len(lowerCAmelCase_ ) + 2 for c in columns] lowerCAmelCase_ : Optional[Any] = max([len(lowerCAmelCase_ ) for name in model_names] ) + 2 # Build the table per se lowerCAmelCase_ : Dict = '''|''' + '''|'''.join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for c, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" lowerCAmelCase_ : List[str] = {True: '''✅''', False: '''❌'''} for name in model_names: lowerCAmelCase_ : List[Any] = model_name_to_prefix[name] lowerCAmelCase_ : Union[str, Any] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for l, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + "|\n" return table def lowerCAmelCase ( lowerCAmelCase_=False )-> Tuple: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = _find_text_in_file( filename=os.path.join(lowerCAmelCase_ , '''index.md''' ) , start_prompt='''<!--This table is updated automatically from the auto modules''' , end_prompt='''<!-- End table-->''' , ) lowerCAmelCase_ : Tuple = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCAmelCase_ , '''index.md''' ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Tuple =parser.parse_args() check_model_table(args.fix_and_overwrite)
262
1
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class snake_case__: '''simple docstring''' def __init__( self , __lowercase , __lowercase=1_4 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=9_9 , __lowercase=3_2 , __lowercase=5 , __lowercase=4 , __lowercase=3_7 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_1_2 , __lowercase=1_6 , __lowercase=2 , __lowercase=0.02 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> Any: lowerCAmelCase_ : List[Any] = parent lowerCAmelCase_ : List[Any] = batch_size lowerCAmelCase_ : List[str] = seq_length lowerCAmelCase_ : List[Any] = is_training lowerCAmelCase_ : Union[str, Any] = use_token_type_ids lowerCAmelCase_ : Optional[int] = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Union[str, Any] = use_mc_token_ids lowerCAmelCase_ : Any = vocab_size lowerCAmelCase_ : Optional[Any] = hidden_size lowerCAmelCase_ : Optional[int] = num_hidden_layers lowerCAmelCase_ : Dict = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : List[str] = hidden_act lowerCAmelCase_ : Dict = hidden_dropout_prob lowerCAmelCase_ : Tuple = attention_probs_dropout_prob lowerCAmelCase_ : Optional[int] = max_position_embeddings lowerCAmelCase_ : Union[str, Any] = type_vocab_size lowerCAmelCase_ : List[str] = type_sequence_label_size lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : Any = num_labels lowerCAmelCase_ : List[Any] = num_choices lowerCAmelCase_ : Optional[int] = scope lowerCAmelCase_ : Optional[int] = self.vocab_size - 1 def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ : Tuple = None if self.use_input_mask: lowerCAmelCase_ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : Dict = None if self.use_token_type_ids: lowerCAmelCase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase_ : Optional[int] = None if self.use_mc_token_ids: lowerCAmelCase_ : Tuple = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) lowerCAmelCase_ : Dict = None lowerCAmelCase_ : int = None lowerCAmelCase_ : Tuple = None if self.use_labels: lowerCAmelCase_ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ : List[str] = self.get_config() lowerCAmelCase_ : Optional[int] = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowercase_ ( self ) -> int: return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase ) -> str: lowerCAmelCase_ : List[Any] = CTRLModel(config=__lowercase ) model.to(__lowercase ) model.eval() model(__lowercase , token_type_ids=__lowercase , head_mask=__lowercase ) model(__lowercase , token_type_ids=__lowercase ) lowerCAmelCase_ : Dict = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase ) -> Dict: lowerCAmelCase_ : Any = CTRLLMHeadModel(__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Any = model(__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() ( ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ) : List[Any] = config_and_inputs lowerCAmelCase_ : str = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask} return config, inputs_dict def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase ) -> List[str]: lowerCAmelCase_ : int = self.num_labels lowerCAmelCase_ : str = CTRLForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() lowerCAmelCase_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : List[str] = model(__lowercase , token_type_ids=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class snake_case__( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Tuple = (CTRLLMHeadModel,) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Dict = ( { """feature-extraction""": CTRLModel, """text-classification""": CTRLForSequenceClassification, """text-generation""": CTRLLMHeadModel, """zero-shot""": CTRLForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : Dict = True SCREAMING_SNAKE_CASE__ : Optional[Any] = False SCREAMING_SNAKE_CASE__ : Optional[int] = False def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[Any]: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : List[Any] = CTRLModelTester(self ) lowerCAmelCase_ : Union[str, Any] = ConfigTester(self , config_class=__lowercase , n_embd=3_7 ) def lowercase_ ( self ) -> int: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> List[Any]: self.config_tester.run_common_tests() def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*__lowercase ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*__lowercase ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def lowercase_ ( self ) -> Any: pass @slow def lowercase_ ( self ) -> str: for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Union[str, Any] = CTRLModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def lowercase_ ( self ) -> Dict: pass @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> Optional[Any]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowercase_ ( self ) -> str: lowerCAmelCase_ : str = CTRLLMHeadModel.from_pretrained('''ctrl''' ) model.to(__lowercase ) lowerCAmelCase_ : int = torch.tensor( [[1_1_8_5_9, 0, 1_6_1_1, 8]] , dtype=torch.long , device=__lowercase ) # Legal the president is lowerCAmelCase_ : str = [ 1_1_8_5_9, 0, 1_6_1_1, 8, 5, 1_5_0, 2_6_4_4_9, 2, 1_9, 3_4_8, 4_6_9, 3, 2_5_9_5, 4_8, 2_0_7_4_0, 2_4_6_5_3_3, 2_4_6_5_3_3, 1_9, 3_0, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a lowerCAmelCase_ : Tuple = model.generate(__lowercase , do_sample=__lowercase ) self.assertListEqual(output_ids[0].tolist() , __lowercase )
262
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
1
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingface_hub.utils import ( EntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, is_jinja_available, ) from packaging import version from requests import HTTPError from .. import __version__ from .constants import ( DEPRECATED_REVISION_ARGS, DIFFUSERS_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, SAFETENSORS_WEIGHTS_NAME, WEIGHTS_NAME, ) from .import_utils import ( ENV_VARS_TRUE_VALUES, _flax_version, _jax_version, _onnxruntime_version, _torch_version, is_flax_available, is_onnx_available, is_torch_available, ) from .logging import get_logger _UpperCAmelCase : int =get_logger(__name__) _UpperCAmelCase : Union[str, Any] =Path(__file__).parent / """model_card_template.md""" _UpperCAmelCase : Optional[Any] =uuida().hex _UpperCAmelCase : Tuple =os.getenv("""HF_HUB_OFFLINE""", """""").upper() in ENV_VARS_TRUE_VALUES _UpperCAmelCase : int =os.getenv("""DISABLE_TELEMETRY""", """""").upper() in ENV_VARS_TRUE_VALUES _UpperCAmelCase : Tuple =HUGGINGFACE_CO_RESOLVE_ENDPOINT + """/api/telemetry/""" def lowerCAmelCase ( lowerCAmelCase_ = None )-> str: lowerCAmelCase_ : Union[str, Any] = f"""diffusers/{__version__}; python/{sys.version.split()[0]}; session_id/{SESSION_ID}""" if DISABLE_TELEMETRY or HF_HUB_OFFLINE: return ua + "; telemetry/off" if is_torch_available(): ua += f"""; torch/{_torch_version}""" if is_flax_available(): ua += f"""; jax/{_jax_version}""" ua += f"""; flax/{_flax_version}""" if is_onnx_available(): ua += f"""; onnxruntime/{_onnxruntime_version}""" # CI will set this value to True if os.environ.get('''DIFFUSERS_IS_CI''' , '''''' ).upper() in ENV_VARS_TRUE_VALUES: ua += "; is_ci/true" if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): ua += "; " + "; ".join(f"""{k}/{v}""" for k, v in user_agent.items() ) elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): ua += "; " + user_agent return ua def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None )-> str: if token is None: lowerCAmelCase_ : int = HfFolder.get_token() if organization is None: lowerCAmelCase_ : Optional[int] = whoami(lowerCAmelCase_ )['''name'''] return f"""{username}/{model_id}""" else: return f"""{organization}/{model_id}""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: if not is_jinja_available(): raise ValueError( '''Modelcard rendering is based on Jinja templates.''' ''' Please make sure to have `jinja` installed before using `create_model_card`.''' ''' To install it, please run `pip install Jinja2`.''' ) if hasattr(lowerCAmelCase_ , '''local_rank''' ) and args.local_rank not in [-1, 0]: return lowerCAmelCase_ : Any = args.hub_token if hasattr(lowerCAmelCase_ , '''hub_token''' ) else None lowerCAmelCase_ : List[Any] = get_full_repo_name(lowerCAmelCase_ , token=lowerCAmelCase_ ) lowerCAmelCase_ : Dict = ModelCard.from_template( card_data=ModelCardData( # Card metadata object that will be converted to YAML block language='''en''' , license='''apache-2.0''' , library_name='''diffusers''' , tags=[] , datasets=args.dataset_name , metrics=[] , ) , template_path=lowerCAmelCase_ , model_name=lowerCAmelCase_ , repo_name=lowerCAmelCase_ , dataset_name=args.dataset_name if hasattr(lowerCAmelCase_ , '''dataset_name''' ) else None , learning_rate=args.learning_rate , train_batch_size=args.train_batch_size , eval_batch_size=args.eval_batch_size , gradient_accumulation_steps=( args.gradient_accumulation_steps if hasattr(lowerCAmelCase_ , '''gradient_accumulation_steps''' ) else None ) , adam_betaa=args.adam_betaa if hasattr(lowerCAmelCase_ , '''adam_beta1''' ) else None , adam_betaa=args.adam_betaa if hasattr(lowerCAmelCase_ , '''adam_beta2''' ) else None , adam_weight_decay=args.adam_weight_decay if hasattr(lowerCAmelCase_ , '''adam_weight_decay''' ) else None , adam_epsilon=args.adam_epsilon if hasattr(lowerCAmelCase_ , '''adam_epsilon''' ) else None , lr_scheduler=args.lr_scheduler if hasattr(lowerCAmelCase_ , '''lr_scheduler''' ) else None , lr_warmup_steps=args.lr_warmup_steps if hasattr(lowerCAmelCase_ , '''lr_warmup_steps''' ) else None , ema_inv_gamma=args.ema_inv_gamma if hasattr(lowerCAmelCase_ , '''ema_inv_gamma''' ) else None , ema_power=args.ema_power if hasattr(lowerCAmelCase_ , '''ema_power''' ) else None , ema_max_decay=args.ema_max_decay if hasattr(lowerCAmelCase_ , '''ema_max_decay''' ) else None , mixed_precision=args.mixed_precision , ) lowerCAmelCase_ : Tuple = os.path.join(args.output_dir , '''README.md''' ) model_card.save(lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = None )-> Union[str, Any]: if resolved_file is None or commit_hash is not None: return commit_hash lowerCAmelCase_ : str = str(Path(lowerCAmelCase_ ).as_posix() ) lowerCAmelCase_ : Dict = re.search(r'''snapshots/([^/]+)/''' , lowerCAmelCase_ ) if search is None: return None lowerCAmelCase_ : Tuple = search.groups()[0] return commit_hash if REGEX_COMMIT_HASH.match(lowerCAmelCase_ ) else None # Old default cache path, potentially to be migrated. # This logic was more or less taken from `transformers`, with the following differences: # - Diffusers doesn't use custom environment variables to specify the cache path. # - There is no need to migrate the cache format, just move the files to the new location. _UpperCAmelCase : Any =os.path.expanduser( os.getenv("""HF_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """huggingface""")) ) _UpperCAmelCase : List[str] =os.path.join(hf_cache_home, """diffusers""") def lowerCAmelCase ( lowerCAmelCase_ = None , lowerCAmelCase_ = None )-> None: if new_cache_dir is None: lowerCAmelCase_ : Dict = DIFFUSERS_CACHE if old_cache_dir is None: lowerCAmelCase_ : Optional[Any] = old_diffusers_cache lowerCAmelCase_ : int = Path(lowerCAmelCase_ ).expanduser() lowerCAmelCase_ : Optional[Any] = Path(lowerCAmelCase_ ).expanduser() for old_blob_path in old_cache_dir.glob('''**/blobs/*''' ): if old_blob_path.is_file() and not old_blob_path.is_symlink(): lowerCAmelCase_ : Dict = new_cache_dir / old_blob_path.relative_to(lowerCAmelCase_ ) new_blob_path.parent.mkdir(parents=lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) os.replace(lowerCAmelCase_ , lowerCAmelCase_ ) try: os.symlink(lowerCAmelCase_ , lowerCAmelCase_ ) except OSError: logger.warning( '''Could not create symlink between old cache and new cache. If you use an older version of diffusers again, files will be re-downloaded.''' ) # At this point, old_cache_dir contains symlinks to the new cache (it can still be used). _UpperCAmelCase : List[str] =os.path.join(DIFFUSERS_CACHE, """version_diffusers_cache.txt""") if not os.path.isfile(cache_version_file): _UpperCAmelCase : Any =0 else: with open(cache_version_file) as f: try: _UpperCAmelCase : Optional[int] =int(f.read()) except ValueError: _UpperCAmelCase : List[Any] =0 if cache_version < 1: _UpperCAmelCase : Optional[Any] =os.path.isdir(old_diffusers_cache) and len(os.listdir(old_diffusers_cache)) > 0 if old_cache_is_not_empty: logger.warning( """The cache for model files in Diffusers v0.14.0 has moved to a new location. Moving your """ """existing cached models. This is a one-time operation, you can interrupt it or run it """ """later by calling `diffusers.utils.hub_utils.move_cache()`.""" ) try: move_cache() except Exception as e: _UpperCAmelCase : str ="""\n""".join(traceback.format_tb(e.__traceback__)) logger.error( f"""There was a problem when trying to move your cache:\n\n{trace}\n{e.__class__.__name__}: {e}\n\nPlease """ """file an issue at https://github.com/huggingface/diffusers/issues/new/choose, copy paste this whole """ """message and we will do our best to help.""" ) if cache_version < 1: try: os.makedirs(DIFFUSERS_CACHE, exist_ok=True) with open(cache_version_file, """w""") as f: f.write("""1""") except Exception: logger.warning( f"""There was a problem when trying to write in your cache folder ({DIFFUSERS_CACHE}). Please, ensure """ """the directory exists and can be written to.""" ) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = None )-> str: if variant is not None: lowerCAmelCase_ : Dict = weights_name.split('''.''' ) lowerCAmelCase_ : Optional[int] = splits[:-1] + [variant] + splits[-1:] lowerCAmelCase_ : Union[str, Any] = '''.'''.join(lowerCAmelCase_ ) return weights_name def lowerCAmelCase ( lowerCAmelCase_ , *, lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , )-> Dict: lowerCAmelCase_ : Optional[Any] = str(lowerCAmelCase_ ) if os.path.isfile(lowerCAmelCase_ ): return pretrained_model_name_or_path elif os.path.isdir(lowerCAmelCase_ ): if os.path.isfile(os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) ): # Load from a PyTorch checkpoint lowerCAmelCase_ : Any = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) return model_file elif subfolder is not None and os.path.isfile( os.path.join(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ): lowerCAmelCase_ : str = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) return model_file else: raise EnvironmentError( f"""Error no file named {weights_name} found in directory {pretrained_model_name_or_path}.""" ) else: # 1. First check if deprecated way of loading from branches is used if ( revision in DEPRECATED_REVISION_ARGS and (weights_name == WEIGHTS_NAME or weights_name == SAFETENSORS_WEIGHTS_NAME) and version.parse(version.parse(lowerCAmelCase_ ).base_version ) >= version.parse('''0.20.0''' ) ): try: lowerCAmelCase_ : Union[str, Any] = hf_hub_download( lowerCAmelCase_ , filename=_add_variant(lowerCAmelCase_ , lowerCAmelCase_ ) , cache_dir=lowerCAmelCase_ , force_download=lowerCAmelCase_ , proxies=lowerCAmelCase_ , resume_download=lowerCAmelCase_ , local_files_only=lowerCAmelCase_ , use_auth_token=lowerCAmelCase_ , user_agent=lowerCAmelCase_ , subfolder=lowerCAmelCase_ , revision=revision or commit_hash , ) warnings.warn( f"""Loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'` is deprecated. Loading instead from `revision='main'` with `variant={revision}`. Loading model variants via `revision='{revision}'` will be removed in diffusers v1. Please use `variant='{revision}'` instead.""" , lowerCAmelCase_ , ) return model_file except: # noqa: E722 warnings.warn( f"""You are loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'`. This behavior is deprecated and will be removed in diffusers v1. One should use `variant='{revision}'` instead. However, it appears that {pretrained_model_name_or_path} currently does not have a {_add_variant(lowerCAmelCase_ , lowerCAmelCase_ )} file in the 'main' branch of {pretrained_model_name_or_path}. \n The Diffusers team and community would be very grateful if you could open an issue: https://github.com/huggingface/diffusers/issues/new with the title '{pretrained_model_name_or_path} is missing {_add_variant(lowerCAmelCase_ , lowerCAmelCase_ )}' so that the correct variant file can be added.""" , lowerCAmelCase_ , ) try: # 2. Load model file as usual lowerCAmelCase_ : Dict = hf_hub_download( lowerCAmelCase_ , filename=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , force_download=lowerCAmelCase_ , proxies=lowerCAmelCase_ , resume_download=lowerCAmelCase_ , local_files_only=lowerCAmelCase_ , use_auth_token=lowerCAmelCase_ , user_agent=lowerCAmelCase_ , subfolder=lowerCAmelCase_ , revision=revision or commit_hash , ) return model_file except RepositoryNotFoundError: raise EnvironmentError( f"""{pretrained_model_name_or_path} is not a local folder and is not a valid model identifier """ '''listed on \'https://huggingface.co/models\'\nIf this is a private repository, make sure to pass a ''' '''token having permission to this repo with `use_auth_token` or log in with `huggingface-cli ''' '''login`.''' ) except RevisionNotFoundError: raise EnvironmentError( f"""{revision} is not a valid git identifier (branch name, tag name or commit id) that exists for """ '''this model name. Check the model page at ''' f"""'https://huggingface.co/{pretrained_model_name_or_path}' for available revisions.""" ) except EntryNotFoundError: raise EnvironmentError( f"""{pretrained_model_name_or_path} does not appear to have a file named {weights_name}.""" ) except HTTPError as err: raise EnvironmentError( f"""There was a specific connection error when trying to load {pretrained_model_name_or_path}:\n{err}""" ) except ValueError: raise EnvironmentError( f"""We couldn't connect to '{HUGGINGFACE_CO_RESOLVE_ENDPOINT}' to load this model, couldn't find it""" f""" in the cached files and it looks like {pretrained_model_name_or_path} is not the path to a""" f""" directory containing a file named {weights_name} or""" ''' \nCheckout your internet connection or see how to run the library in''' ''' offline mode at \'https://huggingface.co/docs/diffusers/installation#offline-mode\'.''' ) except EnvironmentError: raise EnvironmentError( f"""Can't load the model for '{pretrained_model_name_or_path}'. If you were trying to load it from """ '''\'https://huggingface.co/models\', make sure you don\'t have a local directory with the same name. ''' f"""Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a directory """ f"""containing a file named {weights_name}""" )
262
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : Tuple =None _UpperCAmelCase : int =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Any ={ """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """facebook/nllb-large-en-ro""": 1024, """facebook/nllb-200-distilled-600M""": 1024, } # fmt: off _UpperCAmelCase : Any =["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : int = NllbTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=False , **__lowercase , ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : int = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : List[Any] = legacy_behaviour super().__init__( vocab_file=__lowercase , tokenizer_file=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , additional_special_tokens=__lowercase , legacy_behaviour=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_file lowerCAmelCase_ : List[Any] = False if not self.vocab_file else True lowerCAmelCase_ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) lowerCAmelCase_ : Optional[Any] = { lang_code: self.convert_tokens_to_ids(__lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase_ : Any = src_lang if src_lang is not None else '''eng_Latn''' lowerCAmelCase_ : str = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase_ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowercase_ ( self ) -> str: return self._src_lang @src_lang.setter def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : int = self(__lowercase , add_special_tokens=__lowercase , return_tensors=__lowercase , **__lowercase ) lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) lowerCAmelCase_ : List[Any] = tgt_lang_id return inputs def lowercase_ ( self , __lowercase , __lowercase = "eng_Latn" , __lowercase = None , __lowercase = "fra_Latn" , **__lowercase , ) -> BatchEncoding: lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : List[str] = tgt_lang return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def lowercase_ ( self ) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[str] = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[str] = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Optional[int] = [self.cur_lang_code] lowerCAmelCase_ : List[Any] = [self.eos_token_id] lowerCAmelCase_ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Any = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Any = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Any = [self.cur_lang_code] lowerCAmelCase_ : Any = [self.eos_token_id] lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ): copyfile(self.vocab_file , __lowercase ) return (out_vocab_file,)
262
1
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor _UpperCAmelCase : int =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PoolFormerImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _UpperCAmelCase : Optional[Any] =NewType("""DataClass""", Any) _UpperCAmelCase : Dict =NewType("""DataClassType""", Any) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase ( lowerCAmelCase_ )-> Callable[[str], Any]: lowerCAmelCase_ : str = {str(lowerCAmelCase_ ): choice for choice in choices} return lambda lowerCAmelCase_ : str_to_choice.get(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( *, lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = None , **lowerCAmelCase_ , )-> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase_ : Dict = {} if aliases is not None: lowerCAmelCase_ : str = aliases if help is not None: lowerCAmelCase_ : Tuple = help return dataclasses.field(metadata=lowerCAmelCase_ , default=lowerCAmelCase_ , default_factory=lowerCAmelCase_ , **lowerCAmelCase_ ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Iterable[DataClassType] def __init__( self , __lowercase , **__lowercase ) -> List[str]: # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase_ : Optional[int] = ArgumentDefaultsHelpFormatter super().__init__(**__lowercase ) if dataclasses.is_dataclass(__lowercase ): lowerCAmelCase_ : Union[str, Any] = [dataclass_types] lowerCAmelCase_ : List[Any] = list(__lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowercase ) @staticmethod def lowercase_ ( __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[Any] = f"""--{field.name}""" lowerCAmelCase_ : Tuple = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowercase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase_ : List[str] = kwargs.pop('''aliases''' , [] ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[Any] = [aliases] lowerCAmelCase_ : Any = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(__lowercase , '''UnionType''' ) and isinstance(__lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowercase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(__lowercase ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase_ : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase_ : Dict = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase_ : str = ( field.type.__args__[0] if isinstance(__lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase_ : List[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase_ : Dict = {} if origin_type is Literal or (isinstance(field.type , __lowercase ) and issubclass(field.type , __lowercase )): if origin_type is Literal: lowerCAmelCase_ : Optional[Any] = field.type.__args__ else: lowerCAmelCase_ : int = [x.value for x in field.type] lowerCAmelCase_ : str = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase_ : str = field.default else: lowerCAmelCase_ : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase_ : Tuple = copy(__lowercase ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase_ : Dict = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase_ : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase_ : int = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase_ : List[Any] = True elif isclass(__lowercase ) and issubclass(__lowercase , __lowercase ): lowerCAmelCase_ : Union[str, Any] = field.type.__args__[0] lowerCAmelCase_ : Dict = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : Any = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase_ : Dict = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : List[Any] = field.default_factory() else: lowerCAmelCase_ : int = True parser.add_argument(__lowercase , *__lowercase , **__lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase_ : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: if hasattr(__lowercase , '''_argument_group_name''' ): lowerCAmelCase_ : str = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase_ : Dict = self try: lowerCAmelCase_ : Dict[str, type] = get_type_hints(__lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(__lowercase ): lowerCAmelCase_ : Any = '''.'''.join(map(__lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(__lowercase ): if not field.init: continue lowerCAmelCase_ : Optional[int] = type_hints[field.name] self._parse_dataclass_field(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase=None , __lowercase=False , __lowercase=True , __lowercase=None , __lowercase=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase_ : str = [] if args_filename: args_files.append(Path(__lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase_ : str = ArgumentParser() args_file_parser.add_argument(__lowercase , type=__lowercase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = args_file_parser.parse_known_args(args=__lowercase ) lowerCAmelCase_ : int = vars(__lowercase ).get(args_file_flag.lstrip('''-''' ) , __lowercase ) if cmd_args_file_paths: args_files.extend([Path(__lowercase ) for p in cmd_args_file_paths] ) lowerCAmelCase_ : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.parse_known_args(args=__lowercase ) lowerCAmelCase_ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase_ : str = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : str = {k: v for k, v in vars(__lowercase ).items() if k in keys} for k in keys: delattr(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = dtype(**__lowercase ) outputs.append(__lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : int = set(args.keys() ) lowerCAmelCase_ : str = [] for dtype in self.dataclass_types: lowerCAmelCase_ : int = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : List[str] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase_ : List[str] = dtype(**__lowercase ) outputs.append(__lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(__lowercase )}""" ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: with open(Path(__lowercase ) , encoding='''utf-8''' ) as open_json_file: lowerCAmelCase_ : Dict = json.loads(open_json_file.read() ) lowerCAmelCase_ : str = self.parse_dict(__lowercase , allow_extra_keys=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : Optional[Any] = self.parse_dict(yaml.safe_load(Path(__lowercase ).read_text() ) , allow_extra_keys=__lowercase ) return tuple(__lowercase )
262
1
from copy import deepcopy class snake_case__: '''simple docstring''' def __init__( self , __lowercase = None , __lowercase = None ) -> None: if arr is None and size is not None: lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : List[Any] = [0] * size elif arr is not None: self.init(__lowercase ) else: raise ValueError('''Either arr or size must be specified''' ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[Any] = len(__lowercase ) lowerCAmelCase_ : Optional[Any] = deepcopy(__lowercase ) for i in range(1 , self.size ): lowerCAmelCase_ : Any = self.next_(__lowercase ) if j < self.size: self.tree[j] += self.tree[i] def lowercase_ ( self ) -> list[int]: lowerCAmelCase_ : Union[str, Any] = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): lowerCAmelCase_ : List[Any] = self.next_(__lowercase ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowercase_ ( __lowercase ) -> int: return index + (index & (-index)) @staticmethod def lowercase_ ( __lowercase ) -> int: return index - (index & (-index)) def lowercase_ ( self , __lowercase , __lowercase ) -> None: if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value lowerCAmelCase_ : List[Any] = self.next_(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase ) -> None: self.add(__lowercase , value - self.get(__lowercase ) ) def lowercase_ ( self , __lowercase ) -> int: if right == 0: return 0 lowerCAmelCase_ : List[str] = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] lowerCAmelCase_ : str = self.prev(__lowercase ) return result def lowercase_ ( self , __lowercase , __lowercase ) -> int: return self.prefix(__lowercase ) - self.prefix(__lowercase ) def lowercase_ ( self , __lowercase ) -> int: return self.query(__lowercase , index + 1 ) def lowercase_ ( self , __lowercase ) -> int: value -= self.tree[0] if value < 0: return -1 lowerCAmelCase_ : Dict = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 lowerCAmelCase_ : str = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
262
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: return EnvironmentCommand() class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @staticmethod def lowercase_ ( __lowercase ) -> List[Any]: lowerCAmelCase_ : List[str] = parser.add_parser('''env''' ) download_parser.set_defaults(func=__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = huggingface_hub.__version__ lowerCAmelCase_ : str = '''not installed''' lowerCAmelCase_ : str = '''NA''' if is_torch_available(): import torch lowerCAmelCase_ : Any = torch.__version__ lowerCAmelCase_ : str = torch.cuda.is_available() lowerCAmelCase_ : List[str] = '''not installed''' if is_transformers_available(): import transformers lowerCAmelCase_ : Any = transformers.__version__ lowerCAmelCase_ : Optional[Any] = '''not installed''' if is_accelerate_available(): import accelerate lowerCAmelCase_ : List[Any] = accelerate.__version__ lowerCAmelCase_ : List[str] = '''not installed''' if is_xformers_available(): import xformers lowerCAmelCase_ : Optional[Any] = xformers.__version__ lowerCAmelCase_ : int = { '''`diffusers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Huggingface_hub version''': hub_version, '''Transformers version''': transformers_version, '''Accelerate version''': accelerate_version, '''xFormers version''': xformers_version, '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def lowercase_ ( __lowercase ) -> str: return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
262
1
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = JukeboxTokenizer SCREAMING_SNAKE_CASE__ : int = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowercase_ ( self ) -> Union[str, Any]: import torch lowerCAmelCase_ : Union[str, Any] = JukeboxTokenizer.from_pretrained('''openai/jukebox-1b-lyrics''' ) lowerCAmelCase_ : Any = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase_ ( self ) -> List[Any]: import torch lowerCAmelCase_ : Any = JukeboxTokenizer.from_pretrained('''openai/jukebox-5b-lyrics''' ) lowerCAmelCase_ : str = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : Tuple = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
262
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = JukeboxTokenizer SCREAMING_SNAKE_CASE__ : int = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowercase_ ( self ) -> Union[str, Any]: import torch lowerCAmelCase_ : Union[str, Any] = JukeboxTokenizer.from_pretrained('''openai/jukebox-1b-lyrics''' ) lowerCAmelCase_ : Any = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase_ ( self ) -> List[Any]: import torch lowerCAmelCase_ : Any = JukeboxTokenizer.from_pretrained('''openai/jukebox-5b-lyrics''' ) lowerCAmelCase_ : str = tokenizer(**self.metas )['''input_ids'''] # fmt: off lowerCAmelCase_ : Tuple = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
262
1
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class snake_case__( UpperCAmelCase__, UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = AutoencoderKL SCREAMING_SNAKE_CASE__ : Dict = """sample""" SCREAMING_SNAKE_CASE__ : List[str] = 1e-2 @property def lowercase_ ( self ) -> str: lowerCAmelCase_ : Dict = 4 lowerCAmelCase_ : List[Any] = 3 lowerCAmelCase_ : str = (3_2, 3_2) lowerCAmelCase_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes ).to(__lowercase ) return {"sample": image} @property def lowercase_ ( self ) -> Tuple: return (3, 3_2, 3_2) @property def lowercase_ ( self ) -> Optional[Any]: return (3, 3_2, 3_2) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Dict = { '''block_out_channels''': [3_2, 6_4], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } lowerCAmelCase_ : Dict = self.dummy_input return init_dict, inputs_dict def lowercase_ ( self ) -> List[Any]: pass def lowercase_ ( self ) -> List[Any]: pass @unittest.skipIf(torch_device == '''mps''' , '''Gradient checkpointing skipped on MPS''' ) def lowercase_ ( self ) -> Dict: # enable deterministic behavior for gradient checkpointing lowerCAmelCase_ , lowerCAmelCase_ : int = self.prepare_init_args_and_inputs_for_common() lowerCAmelCase_ : Dict = self.model_class(**__lowercase ) model.to(__lowercase ) assert not model.is_gradient_checkpointing and model.training lowerCAmelCase_ : Dict = model(**__lowercase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() lowerCAmelCase_ : List[Any] = torch.randn_like(__lowercase ) lowerCAmelCase_ : List[str] = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing lowerCAmelCase_ : Any = self.model_class(**__lowercase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(__lowercase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training lowerCAmelCase_ : Optional[Any] = model_a(**__lowercase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() lowerCAmelCase_ : Any = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1e-5 ) lowerCAmelCase_ : Any = dict(model.named_parameters() ) lowerCAmelCase_ : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5e-5 ) ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''' , output_loading_info=__lowercase ) self.assertIsNotNone(__lowercase ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(__lowercase ) lowerCAmelCase_ : Tuple = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Optional[Any] = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''' ) lowerCAmelCase_ : Optional[int] = model.to(__lowercase ) model.eval() if torch_device == "mps": lowerCAmelCase_ : Optional[Any] = torch.manual_seed(0 ) else: lowerCAmelCase_ : Tuple = torch.Generator(device=__lowercase ).manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) lowerCAmelCase_ : Any = image.to(__lowercase ) with torch.no_grad(): lowerCAmelCase_ : List[Any] = model(__lowercase , sample_posterior=__lowercase , generator=__lowercase ).sample lowerCAmelCase_ : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": lowerCAmelCase_ : Optional[Any] = torch.tensor( [ -4.0_078e-01, -3.8_323e-04, -1.2_681e-01, -1.1_462e-01, 2.0_095e-01, 1.0_893e-01, -8.8_247e-02, -3.0_361e-01, -9.8_644e-03, ] ) elif torch_device == "cpu": lowerCAmelCase_ : str = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: lowerCAmelCase_ : Dict = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(__lowercase , __lowercase , rtol=1e-2 ) ) @slow class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self , __lowercase , __lowercase ) -> Optional[Any]: return f"""gaussian_noise_s={seed}_shape={"_".join([str(__lowercase ) for s in shape] )}.npy""" def lowercase_ ( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self , __lowercase=0 , __lowercase=(4, 3, 5_1_2, 5_1_2) , __lowercase=False ) -> Any: lowerCAmelCase_ : Optional[int] = torch.floataa if fpaa else torch.floataa lowerCAmelCase_ : List[Any] = torch.from_numpy(load_hf_numpy(self.get_file_format(__lowercase , __lowercase ) ) ).to(__lowercase ).to(__lowercase ) return image def lowercase_ ( self , __lowercase="CompVis/stable-diffusion-v1-4" , __lowercase=False ) -> str: lowerCAmelCase_ : Union[str, Any] = '''fp16''' if fpaa else None lowerCAmelCase_ : Union[str, Any] = torch.floataa if fpaa else torch.floataa lowerCAmelCase_ : str = AutoencoderKL.from_pretrained( __lowercase , subfolder='''vae''' , torch_dtype=__lowercase , revision=__lowercase , ) model.to(__lowercase ).eval() return model def lowercase_ ( self , __lowercase=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(__lowercase ) return torch.Generator(device=__lowercase ).manual_seed(__lowercase ) @parameterized.expand( [ # fmt: off [3_3, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [4_7, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> int: lowerCAmelCase_ : Dict = self.get_sd_vae_model() lowerCAmelCase_ : Any = self.get_sd_image(__lowercase ) lowerCAmelCase_ : Optional[Any] = self.get_generator(__lowercase ) with torch.no_grad(): lowerCAmelCase_ : List[str] = model(__lowercase , generator=__lowercase , sample_posterior=__lowercase ).sample assert sample.shape == image.shape lowerCAmelCase_ : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() lowerCAmelCase_ : Optional[Any] = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice ) assert torch_all_close(__lowercase , __lowercase , atol=3e-3 ) @parameterized.expand( [ # fmt: off [3_3, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [4_7, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def lowercase_ ( self , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = self.get_sd_vae_model(fpaa=__lowercase ) lowerCAmelCase_ : int = self.get_sd_image(__lowercase , fpaa=__lowercase ) lowerCAmelCase_ : List[str] = self.get_generator(__lowercase ) with torch.no_grad(): lowerCAmelCase_ : List[str] = model(__lowercase , generator=__lowercase , sample_posterior=__lowercase ).sample assert sample.shape == image.shape lowerCAmelCase_ : int = sample[-1, -2:, :2, -2:].flatten().float().cpu() lowerCAmelCase_ : List[str] = torch.tensor(__lowercase ) assert torch_all_close(__lowercase , __lowercase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [4_7, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Optional[int] = self.get_sd_vae_model() lowerCAmelCase_ : Union[str, Any] = self.get_sd_image(__lowercase ) with torch.no_grad(): lowerCAmelCase_ : List[Any] = model(__lowercase ).sample assert sample.shape == image.shape lowerCAmelCase_ : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() lowerCAmelCase_ : str = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice ) assert torch_all_close(__lowercase , __lowercase , atol=3e-3 ) @parameterized.expand( [ # fmt: off [1_3, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [3_7, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def lowercase_ ( self , __lowercase , __lowercase ) -> int: lowerCAmelCase_ : int = self.get_sd_vae_model() lowerCAmelCase_ : Tuple = self.get_sd_image(__lowercase , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): lowerCAmelCase_ : str = model.decode(__lowercase ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] lowerCAmelCase_ : str = sample[-1, -2:, :2, -2:].flatten().cpu() lowerCAmelCase_ : Union[str, Any] = torch.tensor(__lowercase ) assert torch_all_close(__lowercase , __lowercase , atol=1e-3 ) @parameterized.expand( [ # fmt: off [2_7, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [1_6, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def lowercase_ ( self , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Any = self.get_sd_vae_model(fpaa=__lowercase ) lowerCAmelCase_ : Any = self.get_sd_image(__lowercase , shape=(3, 4, 6_4, 6_4) , fpaa=__lowercase ) with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model.decode(__lowercase ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] lowerCAmelCase_ : Tuple = sample[-1, -2:, :2, -2:].flatten().float().cpu() lowerCAmelCase_ : Tuple = torch.tensor(__lowercase ) assert torch_all_close(__lowercase , __lowercase , atol=5e-3 ) @parameterized.expand([(1_3,), (1_6,), (2_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''' ) def lowercase_ ( self , __lowercase ) -> Optional[Any]: lowerCAmelCase_ : List[Any] = self.get_sd_vae_model(fpaa=__lowercase ) lowerCAmelCase_ : List[Any] = self.get_sd_image(__lowercase , shape=(3, 4, 6_4, 6_4) , fpaa=__lowercase ) with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model.decode(__lowercase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): lowerCAmelCase_ : int = model.decode(__lowercase ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(__lowercase , __lowercase , atol=1e-1 ) @parameterized.expand([(1_3,), (1_6,), (3_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''' ) def lowercase_ ( self , __lowercase ) -> Optional[int]: lowerCAmelCase_ : Any = self.get_sd_vae_model() lowerCAmelCase_ : List[Any] = self.get_sd_image(__lowercase , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): lowerCAmelCase_ : List[Any] = model.decode(__lowercase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): lowerCAmelCase_ : Dict = model.decode(__lowercase ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(__lowercase , __lowercase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [4_7, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def lowercase_ ( self , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Any = self.get_sd_vae_model() lowerCAmelCase_ : Any = self.get_sd_image(__lowercase ) lowerCAmelCase_ : Dict = self.get_generator(__lowercase ) with torch.no_grad(): lowerCAmelCase_ : Tuple = model.encode(__lowercase ).latent_dist lowerCAmelCase_ : str = dist.sample(generator=__lowercase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] lowerCAmelCase_ : int = sample[0, -1, -3:, -3:].flatten().cpu() lowerCAmelCase_ : int = torch.tensor(__lowercase ) lowerCAmelCase_ : List[Any] = 3e-3 if torch_device != '''mps''' else 1e-2 assert torch_all_close(__lowercase , __lowercase , atol=__lowercase )
262
from __future__ import annotations import requests def lowerCAmelCase ( lowerCAmelCase_ )-> dict: lowerCAmelCase_ : List[Any] = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> list[dict]: lowerCAmelCase_ : List[Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCAmelCase_ : Tuple = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def lowerCAmelCase ( lowerCAmelCase_ = 10 )-> str: lowerCAmelCase_ : Optional[Any] = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join('''* [{title}]({url})'''.format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
262
1
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
262
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : List[str] =get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCAmelCase : Optional[int] =25_0004 _UpperCAmelCase : Tuple =25_0020 @require_sentencepiece @require_tokenizers class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = MBartTokenizer SCREAMING_SNAKE_CASE__ : Dict = MBartTokenizerFast SCREAMING_SNAKE_CASE__ : Tuple = True SCREAMING_SNAKE_CASE__ : List[str] = True def lowercase_ ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : str = MBartTokenizer(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[int] = MBartTokenizer(__lowercase , keep_accents=__lowercase ) lowerCAmelCase_ : Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) lowerCAmelCase_ : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCAmelCase_ : Dict = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowercase_ ( self ) -> Dict: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase_ : int = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase_ : Optional[int] = self.rust_tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : int = self.tokenizer_class.from_pretrained(__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = tempfile.mkdtemp() lowerCAmelCase_ : Union[str, Any] = tokenizer_r.save_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase_ : str = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Tuple = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : Dict = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : int = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Tuple = tokenizer_p.save_pretrained(__lowercase ) # Checks it save with the same files self.assertSequenceEqual(__lowercase , __lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Optional[int] = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase_ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = tokenizer_r.save_pretrained(__lowercase , legacy_format=__lowercase ) lowerCAmelCase_ : Optional[int] = tokenizer_p.save_pretrained(__lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase_ : Dict = tokenizer_r.from_pretrained(__lowercase ) lowerCAmelCase_ : List[Any] = tokenizer_p.from_pretrained(__lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowercase , __lowercase ) ) shutil.rmtree(__lowercase ) @require_torch @require_sentencepiece @require_tokenizers class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = """facebook/mbart-large-en-ro""" SCREAMING_SNAKE_CASE__ : int = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] SCREAMING_SNAKE_CASE__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] SCREAMING_SNAKE_CASE__ : str = [8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2, EN_CODE] @classmethod def lowercase_ ( cls ) -> Optional[int]: lowerCAmelCase_ : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) lowerCAmelCase_ : Optional[Any] = 1 return cls def lowercase_ ( self ) -> Optional[Any]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 2_5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 2_5_0_0_0_4 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 2_5_0_0_2_0 ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Optional[int] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) def lowercase_ ( self ) -> Any: self.assertIn(__lowercase , self.tokenizer.all_special_ids ) lowerCAmelCase_ : Union[str, Any] = [RO_CODE, 8_8_4, 9_0_1_9, 9_6, 9, 9_1_6, 8_6_7_9_2, 3_6, 1_8_7_4_3, 1_5_5_9_6, 5, 2] lowerCAmelCase_ : Tuple = self.tokenizer.decode(__lowercase , skip_special_tokens=__lowercase ) lowerCAmelCase_ : List[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowercase ) self.assertEqual(__lowercase , __lowercase ) self.assertNotIn(self.tokenizer.eos_token , __lowercase ) def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Union[str, Any] = ['''this is gunna be a long sentence ''' * 2_0] assert isinstance(src_text[0] , __lowercase ) lowerCAmelCase_ : str = 1_0 lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , max_length=__lowercase , truncation=__lowercase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __lowercase ) self.assertEqual(len(__lowercase ) , __lowercase ) def lowercase_ ( self ) -> int: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [2_5_0_0_2_6, 2_5_0_0_0_1] ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = tempfile.mkdtemp() lowerCAmelCase_ : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = MBartTokenizer.from_pretrained(__lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowercase ) @require_torch def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Tuple = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowercase , return_tensors='''pt''' ) lowerCAmelCase_ : Tuple = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : str = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) lowerCAmelCase_ : int = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(__lowercase , __lowercase ) self.assertEqual((2, 1_4) , batch.input_ids.shape ) self.assertEqual((2, 1_4) , batch.attention_mask.shape ) lowerCAmelCase_ : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Optional[Any] = self.tokenizer(self.src_text , padding=__lowercase , truncation=__lowercase , max_length=3 , return_tensors='''pt''' ) lowerCAmelCase_ : Any = self.tokenizer( text_target=self.tgt_text , padding=__lowercase , truncation=__lowercase , max_length=1_0 , return_tensors='''pt''' ) lowerCAmelCase_ : int = targets['''input_ids'''] lowerCAmelCase_ : Optional[Any] = shift_tokens_right(__lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(__lowercase ) , { # A, test, EOS, en_XX '''input_ids''': [[6_2, 3_0_3_4, 2, 2_5_0_0_0_4]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 2_5_0_0_0_1, } , )
262
1
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) lowerCAmelCase_ : Optional[Any] = Vector() def lowercase_ ( self ) -> None: lowerCAmelCase_ : List[str] = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(__lowercase ) , '''(0,0,0,0,0,1)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : int = Vector([1, 2, 3, 4] ) self.assertEqual(len(__lowercase ) , 4 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[Any] = Vector([1, 2] ) lowerCAmelCase_ : Dict = Vector([1, 2, 3, 4, 5] ) lowerCAmelCase_ : Tuple = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) lowerCAmelCase_ : Dict = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.2_36 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.4_16 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.6_16 , 3 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Vector([1, 2, 3] ) lowerCAmelCase_ : Optional[Any] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[Any] = Vector([1, 2, 3] ) lowerCAmelCase_ : int = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Any = Vector([1, 2, 3] ) lowerCAmelCase_ : Union[str, Any] = Vector([2, -1, 4] ) # for test of dot product lowerCAmelCase_ : List[str] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def lowercase_ ( self ) -> None: self.assertEqual(str(zero_vector(1_0 ) ).count('''0''' ) , 1_0 ) def lowercase_ ( self ) -> None: self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Dict = Vector([1, 2, 3] ) lowerCAmelCase_ : Tuple = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , __lowercase , __lowercase ) ) , '''(3,4,7)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 0, 0, 0, 0, 0] ) lowerCAmelCase_ : List[str] = x.copy() self.assertEqual(str(__lowercase ) , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(__lowercase ) , '''(0,1,0)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : int = [[-3, -1_4, -1_0], [-5, -1_0, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(__lowercase , __lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : str = [[-3, 1_4, -1_0], [5, -1_0, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(__lowercase , __lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Any = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) lowerCAmelCase_ : Dict = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def lowercase_ ( self ) -> None: self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline _UpperCAmelCase : Any =logging.get_logger(__name__) # pylint: disable=invalid-name class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase ) -> Optional[int]: super().__init__() self.register_modules(unet=__lowercase , scheduler=__lowercase ) @torch.no_grad() def __call__( self , __lowercase = 1 , __lowercase = 1_0_0 , __lowercase = None , __lowercase = None , __lowercase = True , ) -> Union[AudioPipelineOutput, Tuple]: if audio_length_in_s is None: lowerCAmelCase_ : Tuple = self.unet.config.sample_size / self.unet.config.sample_rate lowerCAmelCase_ : Any = audio_length_in_s * self.unet.config.sample_rate lowerCAmelCase_ : Any = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( f"""{audio_length_in_s} is too small. Make sure it's bigger or equal to""" f""" {3 * down_scale_factor / self.unet.config.sample_rate}.""" ) lowerCAmelCase_ : List[Any] = int(__lowercase ) if sample_size % down_scale_factor != 0: lowerCAmelCase_ : Optional[Any] = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( f"""{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled""" f""" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising""" ''' process.''' ) lowerCAmelCase_ : int = int(__lowercase ) lowerCAmelCase_ : Dict = next(iter(self.unet.parameters() ) ).dtype lowerCAmelCase_ : Optional[int] = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(__lowercase , __lowercase ) and len(__lowercase ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(__lowercase )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) lowerCAmelCase_ : int = randn_tensor(__lowercase , generator=__lowercase , device=self.device , dtype=__lowercase ) # set step values self.scheduler.set_timesteps(__lowercase , device=audio.device ) lowerCAmelCase_ : Union[str, Any] = self.scheduler.timesteps.to(__lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowerCAmelCase_ : List[Any] = self.unet(__lowercase , __lowercase ).sample # 2. compute previous image: x_t -> t_t-1 lowerCAmelCase_ : List[Any] = self.scheduler.step(__lowercase , __lowercase , __lowercase ).prev_sample lowerCAmelCase_ : Tuple = audio.clamp(-1 , 1 ).float().cpu().numpy() lowerCAmelCase_ : List[str] = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=__lowercase )
262
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
262
1
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> str: pass def lowerCAmelCase ( lowerCAmelCase_ )-> Any: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. _UpperCAmelCase : str =( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> int: lowerCAmelCase_ : List[str] = pipeline( '''document-question-answering''' , model=__lowercase , tokenizer=__lowercase , image_processor=__lowercase ) lowerCAmelCase_ : Optional[int] = INVOICE_URL lowerCAmelCase_ : List[Any] = list(zip(*apply_tesseract(load_image(__lowercase ) , __lowercase , '''''' ) ) ) lowerCAmelCase_ : Optional[Any] = '''What is the placebo?''' lowerCAmelCase_ : Dict = [ { '''image''': load_image(__lowercase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def lowercase_ ( self , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : List[Any] = dqa_pipeline(__lowercase , top_k=2 ) self.assertEqual( __lowercase , [ [ {'''score''': ANY(__lowercase ), '''answer''': ANY(__lowercase ), '''start''': ANY(__lowercase ), '''end''': ANY(__lowercase )}, {'''score''': ANY(__lowercase ), '''answer''': ANY(__lowercase ), '''start''': ANY(__lowercase ), '''end''': ANY(__lowercase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self ) -> str: lowerCAmelCase_ : int = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) lowerCAmelCase_ : Tuple = INVOICE_URL lowerCAmelCase_ : int = '''How many cats are there?''' lowerCAmelCase_ : Dict = [ {'''score''': 0.00_01, '''answer''': '''oy 2312/2019''', '''start''': 3_8, '''end''': 3_9}, {'''score''': 0.00_01, '''answer''': '''oy 2312/2019 DUE''', '''start''': 3_8, '''end''': 4_0}, ] lowerCAmelCase_ : Dict = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual(nested_simplify(__lowercase , decimals=4 ) , __lowercase ) lowerCAmelCase_ : Tuple = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__lowercase , decimals=4 ) , __lowercase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowerCAmelCase_ : Any = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' lowerCAmelCase_ : List[str] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual(__lowercase , [] ) # We can optionnally pass directly the words and bounding boxes lowerCAmelCase_ : int = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' lowerCAmelCase_ : Any = [] lowerCAmelCase_ : Dict = [] lowerCAmelCase_ : List[str] = dqa_pipeline(image=__lowercase , question=__lowercase , words=__lowercase , boxes=__lowercase , top_k=2 ) self.assertEqual(__lowercase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Tuple = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) lowerCAmelCase_ : Dict = INVOICE_URL lowerCAmelCase_ : int = '''What is the invoice number?''' lowerCAmelCase_ : List[Any] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_44, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.00_09, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) lowerCAmelCase_ : Optional[int] = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_44, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.00_09, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) lowerCAmelCase_ : int = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.99_44, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.00_09, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Any = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=5_0 , ) lowerCAmelCase_ : List[Any] = INVOICE_URL lowerCAmelCase_ : str = '''What is the invoice number?''' lowerCAmelCase_ : List[str] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_74, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.99_48, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) lowerCAmelCase_ : Optional[int] = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_74, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.99_48, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) lowerCAmelCase_ : List[str] = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.99_74, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.99_48, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__lowercase ) lowerCAmelCase_ : int = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__lowercase , revision='''3dc6de3''' , ) lowerCAmelCase_ : Optional[int] = INVOICE_URL lowerCAmelCase_ : Dict = '''What is the invoice number?''' lowerCAmelCase_ : Optional[int] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.42_51, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.08_19, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] , ) lowerCAmelCase_ : Tuple = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.42_51, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.08_19, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] , ) lowerCAmelCase_ : Optional[Any] = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.42_51, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.08_19, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] ] * 2 , ) lowerCAmelCase_ : str = list(zip(*apply_tesseract(load_image(__lowercase ) , __lowercase , '''''' ) ) ) # This model should also work if `image` is set to None lowerCAmelCase_ : Any = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.42_51, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.08_19, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Dict = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__lowercase ) lowerCAmelCase_ : Optional[int] = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__lowercase , revision='''3dc6de3''' , max_seq_len=5_0 , ) lowerCAmelCase_ : List[str] = INVOICE_URL lowerCAmelCase_ : Tuple = '''What is the invoice number?''' lowerCAmelCase_ : Optional[Any] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_99, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.99_98, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) lowerCAmelCase_ : Any = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.99_99, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.99_98, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ] * 2 , ) lowerCAmelCase_ : Optional[Any] = list(zip(*apply_tesseract(load_image(__lowercase ) , __lowercase , '''''' ) ) ) # This model should also work if `image` is set to None lowerCAmelCase_ : Union[str, Any] = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.99_99, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.99_98, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] , ) @slow @require_torch def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) lowerCAmelCase_ : List[Any] = INVOICE_URL lowerCAmelCase_ : List[Any] = '''What is the invoice number?''' lowerCAmelCase_ : Optional[Any] = dqa_pipeline(image=__lowercase , question=__lowercase , top_k=2 ) self.assertEqual(nested_simplify(__lowercase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def lowercase_ ( self ) -> Dict: pass
262
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : str ={ """facebook/vit-mae-base""": """https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json""", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """vit_mae""" def __init__( self , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1e-12 , __lowercase=2_2_4 , __lowercase=1_6 , __lowercase=3 , __lowercase=True , __lowercase=1_6 , __lowercase=5_1_2 , __lowercase=8 , __lowercase=2_0_4_8 , __lowercase=0.75 , __lowercase=False , **__lowercase , ) -> str: super().__init__(**__lowercase ) lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : Union[str, Any] = image_size lowerCAmelCase_ : Optional[int] = patch_size lowerCAmelCase_ : Tuple = num_channels lowerCAmelCase_ : List[str] = qkv_bias lowerCAmelCase_ : List[Any] = decoder_num_attention_heads lowerCAmelCase_ : int = decoder_hidden_size lowerCAmelCase_ : Optional[int] = decoder_num_hidden_layers lowerCAmelCase_ : Tuple = decoder_intermediate_size lowerCAmelCase_ : Tuple = mask_ratio lowerCAmelCase_ : Any = norm_pix_loss
262
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = ["""image_processor""", """tokenizer"""] SCREAMING_SNAKE_CASE__ : List[Any] = """LayoutLMv2ImageProcessor""" SCREAMING_SNAKE_CASE__ : int = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self , __lowercase=None , __lowercase=None , **__lowercase ) -> int: if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __lowercase , ) lowerCAmelCase_ : List[str] = kwargs.pop('''feature_extractor''' ) lowerCAmelCase_ : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(__lowercase , __lowercase ) def __call__( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = True , __lowercase = False , __lowercase = None , __lowercase = None , __lowercase = 0 , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = True , __lowercase = None , **__lowercase , ) -> BatchEncoding: # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''' ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''' ) # first, apply the image processor lowerCAmelCase_ : Optional[int] = self.image_processor(images=__lowercase , return_tensors=__lowercase ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[int] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase_ : Any = features['''words'''] lowerCAmelCase_ : Union[str, Any] = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=__lowercase , add_special_tokens=__lowercase , padding=__lowercase , truncation=__lowercase , max_length=__lowercase , stride=__lowercase , pad_to_multiple_of=__lowercase , return_token_type_ids=__lowercase , return_attention_mask=__lowercase , return_overflowing_tokens=__lowercase , return_special_tokens_mask=__lowercase , return_offsets_mapping=__lowercase , return_length=__lowercase , verbose=__lowercase , return_tensors=__lowercase , **__lowercase , ) # add pixel values lowerCAmelCase_ : Any = features.pop('''pixel_values''' ) if return_overflowing_tokens is True: lowerCAmelCase_ : str = self.get_overflowing_images(__lowercase , encoded_inputs['''overflow_to_sample_mapping'''] ) lowerCAmelCase_ : List[str] = images return encoded_inputs def lowercase_ ( self , __lowercase , __lowercase ) -> int: # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image lowerCAmelCase_ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(__lowercase ) != len(__lowercase ): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' f""" {len(__lowercase )} and {len(__lowercase )}""" ) return images_with_overflow def lowercase_ ( self , *__lowercase , **__lowercase ) -> Tuple: return self.tokenizer.batch_decode(*__lowercase , **__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> Any: return self.tokenizer.decode(*__lowercase , **__lowercase ) @property def lowercase_ ( self ) -> Dict: return ["input_ids", "bbox", "attention_mask", "image"] @property def lowercase_ ( self ) -> Tuple: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __lowercase , ) return self.image_processor_class @property def lowercase_ ( self ) -> Any: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __lowercase , ) return self.image_processor
262
def lowerCAmelCase ( lowerCAmelCase_ = 10**9 )-> int: lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : str = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value lowerCAmelCase_ : Any = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f"""{solution() = }""")
262
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _UpperCAmelCase : List[str] =logging.get_logger(__name__) _UpperCAmelCase : str ={"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : str ={ """vocab_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : str ={ """yjernite/retribert-base-uncased""": 512, } _UpperCAmelCase : List[str] ={ """yjernite/retribert-base-uncased""": {"""do_lower_case""": True}, } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE__ : str = RetriBertTokenizer SCREAMING_SNAKE_CASE__ : Any = ["""input_ids""", """attention_mask"""] def __init__( self , __lowercase=None , __lowercase=None , __lowercase=True , __lowercase="[UNK]" , __lowercase="[SEP]" , __lowercase="[PAD]" , __lowercase="[CLS]" , __lowercase="[MASK]" , __lowercase=True , __lowercase=None , **__lowercase , ) -> Union[str, Any]: super().__init__( __lowercase , tokenizer_file=__lowercase , do_lower_case=__lowercase , unk_token=__lowercase , sep_token=__lowercase , pad_token=__lowercase , cls_token=__lowercase , mask_token=__lowercase , tokenize_chinese_chars=__lowercase , strip_accents=__lowercase , **__lowercase , ) lowerCAmelCase_ : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , __lowercase ) != do_lower_case or normalizer_state.get('''strip_accents''' , __lowercase ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , __lowercase ) != tokenize_chinese_chars ): lowerCAmelCase_ : List[str] = getattr(__lowercase , normalizer_state.pop('''type''' ) ) lowerCAmelCase_ : str = do_lower_case lowerCAmelCase_ : Tuple = strip_accents lowerCAmelCase_ : Tuple = tokenize_chinese_chars lowerCAmelCase_ : Tuple = normalizer_class(**__lowercase ) lowerCAmelCase_ : int = do_lower_case def lowercase_ ( self , __lowercase , __lowercase=None ) -> Optional[int]: lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : List[Any] = [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: lowerCAmelCase_ : Any = self._tokenizer.model.save(__lowercase , name=__lowercase ) return tuple(__lowercase )
262
import inspect import unittest class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> int: try: import diffusers # noqa: F401 except ImportError: assert False def lowercase_ ( self ) -> List[str]: import diffusers from diffusers.dependency_versions_table import deps lowerCAmelCase_ : Any = inspect.getmembers(__lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowerCAmelCase_ : Optional[int] = '''k-diffusion''' elif backend == "invisible_watermark": lowerCAmelCase_ : Dict = '''invisible-watermark''' assert backend in deps, f"""{backend} is not in the deps table!"""
262
1
from ..utils import DummyObject, requires_backends class snake_case__( metaclass=UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""torch""", """torchsde"""] def __init__( self , *__lowercase , **__lowercase ) -> List[str]: requires_backends(self , ['''torch''', '''torchsde'''] ) @classmethod def lowercase_ ( cls , *__lowercase , **__lowercase ) -> List[Any]: requires_backends(cls , ['''torch''', '''torchsde'''] ) @classmethod def lowercase_ ( cls , *__lowercase , **__lowercase ) -> Any: requires_backends(cls , ['''torch''', '''torchsde'''] )
262
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCAmelCase : Any =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
1
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class snake_case__: '''simple docstring''' def __init__( self ) -> Any: lowerCAmelCase_ : List[str] = '''''' lowerCAmelCase_ : Any = '''''' lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Tuple = 2_5_6 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : str = 0 def lowercase_ ( self , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = cva.imread(__lowercase , 0 ) lowerCAmelCase_ : List[Any] = copy.deepcopy(self.img ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] , label='''x''' ) lowerCAmelCase_ : Optional[int] = np.sum(__lowercase ) for i in range(len(__lowercase ) ): lowerCAmelCase_ : Tuple = x[i] / self.k self.sk += prk lowerCAmelCase_ : str = (self.L - 1) * self.sk if self.rem != 0: lowerCAmelCase_ : Optional[int] = int(last % last ) lowerCAmelCase_ : Union[str, Any] = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(__lowercase ) lowerCAmelCase_ : int = int(np.ma.count(self.img ) / self.img[1].size ) lowerCAmelCase_ : Optional[int] = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): lowerCAmelCase_ : List[str] = self.img[j][i] if num != self.last_list[num]: lowerCAmelCase_ : Tuple = self.last_list[num] cva.imwrite('''output_data/output.jpg''' , self.img ) def lowercase_ ( self ) -> str: plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] ) def lowercase_ ( self ) -> List[Any]: cva.imshow('''Output-Image''' , self.img ) cva.imshow('''Input-Image''' , self.original_image ) cva.waitKey(5_0_0_0 ) cva.destroyAllWindows() if __name__ == "__main__": _UpperCAmelCase : Tuple =os.path.join(os.path.basename(__file__), """image_data/input.jpg""") _UpperCAmelCase : str =ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
262
def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : Optional[int] = set() # edges = list of graph's edges lowerCAmelCase_ : List[str] = get_edges(lowerCAmelCase_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = edges.pop() chosen_vertices.add(lowerCAmelCase_ ) chosen_vertices.add(lowerCAmelCase_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase_ ) return chosen_vertices def lowerCAmelCase ( lowerCAmelCase_ )-> set: lowerCAmelCase_ : List[Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
262
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCAmelCase : Dict =logging.getLogger(__name__) def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[Any]: return (preds == labels).mean() @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys() )} ) SCREAMING_SNAKE_CASE__ : str = field(metadata={"""help""": """Should contain the data files for the task."""} ) SCREAMING_SNAKE_CASE__ : int = field( default=128, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def lowerCAmelCase ( )-> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Optional[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , lowerCAmelCase_ ) # Set seed set_seed(training_args.seed ) try: lowerCAmelCase_ : List[str] = processors[data_args.task_name]() lowerCAmelCase_ : Dict = processor.get_labels() lowerCAmelCase_ : Optional[Any] = len(lowerCAmelCase_ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase_ : Optional[int] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCAmelCase_ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Optional[Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , ) # Get datasets lowerCAmelCase_ : Optional[Any] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowerCAmelCase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) lowerCAmelCase_ : List[str] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowerCAmelCase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(lowerCAmelCase_ ) -> Dict: lowerCAmelCase_ : Dict = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(lowerCAmelCase_ , p.label_ids )} # Data collator lowerCAmelCase_ : Optional[Any] = DataCollatorWithPadding(lowerCAmelCase_ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowerCAmelCase_ : List[str] = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=lowerCAmelCase_ , eval_dataset=lowerCAmelCase_ , compute_metrics=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowerCAmelCase_ : Tuple = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCAmelCase_ : int = trainer.evaluate() lowerCAmelCase_ : List[str] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(lowerCAmelCase_ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , lowerCAmelCase_ , lowerCAmelCase_ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(lowerCAmelCase_ ) return results def lowerCAmelCase ( lowerCAmelCase_ )-> str: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
from math import sqrt def lowerCAmelCase ( lowerCAmelCase_ )-> bool: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ : Optional[int] = False for divisor in range(2 , int(round(sqrt(lowerCAmelCase_ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ : Tuple = False break # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'status' must been from type bool" return status def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ : Tuple = list(range(2 , n + 1 ) ) lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ : str = 0 # filters actual prime numbers. lowerCAmelCase_ : Optional[int] = [x for x in begin_list if x != 0] # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(lowerCAmelCase_ ): ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ : int = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ : List[Any] = 2 lowerCAmelCase_ : Optional[int] = number if number == 0 or number == 1: ans.append(lowerCAmelCase_ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(lowerCAmelCase_ ): while quotient != 1: if is_prime(lowerCAmelCase_ ) and (quotient % factor == 0): ans.append(lowerCAmelCase_ ) quotient /= factor else: factor += 1 else: ans.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type list" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : Dict = 0 # prime factorization of 'number' lowerCAmelCase_ : Any = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = max(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ : List[Any] = 0 # prime factorization of 'number' lowerCAmelCase_ : Dict = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = min(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'ans' must been from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 == 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 == 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ), "'number' must been an int" assert isinstance(number % 2 != 0 , lowerCAmelCase_ ), "compare bust been from type bool" return number % 2 != 0 def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (number > 2) and is_even(lowerCAmelCase_ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ : str = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ : int = get_prime_numbers(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = len(lowerCAmelCase_ ) # run variable for while-loops. lowerCAmelCase_ : Union[str, Any] = 0 lowerCAmelCase_ : Tuple = None # exit variable. for break up the loops lowerCAmelCase_ : int = True while i < len_pn and loop: lowerCAmelCase_ : int = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ : Tuple = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (len(lowerCAmelCase_ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : int = 0 while numbera != 0: lowerCAmelCase_ : str = numbera % numbera lowerCAmelCase_ : List[Any] = numbera lowerCAmelCase_ : Any = rest # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ : List[Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) lowerCAmelCase_ : int = prime_factorization(lowerCAmelCase_ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = max(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = 0 lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) lowerCAmelCase_ : Tuple = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(max(lowerCAmelCase_ , lowerCAmelCase_ ) ): ans *= n else: lowerCAmelCase_ : List[str] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ : Optional[Any] = prime_fac_a.count(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ): ans *= n done.append(lowerCAmelCase_ ) # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Optional[int] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(lowerCAmelCase_ ): ans += 1 # precondition assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and is_prime( lowerCAmelCase_ ), "'ans' must been a prime number and from type int" return ans def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: assert ( is_prime(lowerCAmelCase_ ) and is_prime(lowerCAmelCase_ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ : Union[str, Any] = p_number_a + 1 # jump to the next number lowerCAmelCase_ : Optional[int] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 while number < p_number_a: ans.append(lowerCAmelCase_ ) number += 1 # fetch the next prime number. while not is_prime(lowerCAmelCase_ ): number += 1 # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ans[0] != p_number_a and ans[len(lowerCAmelCase_ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ : List[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(lowerCAmelCase_ ) # precondition assert ans[0] == 1 and ans[len(lowerCAmelCase_ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCAmelCase ( lowerCAmelCase_ )-> List[Any]: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ : Union[str, Any] = get_divisors(lowerCAmelCase_ ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (divisors[0] == 1) and (divisors[len(lowerCAmelCase_ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Union[str, Any]: assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ : Optional[Any] = gcd(abs(lowerCAmelCase_ ) , abs(lowerCAmelCase_ ) ) # precondition assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCAmelCase ( lowerCAmelCase_ )-> int: assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ : Union[str, Any] = ans ans += fiba lowerCAmelCase_ : Optional[Any] = tmp return ans
262
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCAmelCase : Union[str, Any] =logging.get_logger(__name__) _UpperCAmelCase : Optional[int] ={ """microsoft/resnet-50""": """https://huggingface.co/microsoft/resnet-50/blob/main/config.json""", } class snake_case__( UpperCAmelCase__, UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = """resnet""" SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""basic""", """bottleneck"""] def __init__( self , __lowercase=3 , __lowercase=6_4 , __lowercase=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , __lowercase=[3, 4, 6, 3] , __lowercase="bottleneck" , __lowercase="relu" , __lowercase=False , __lowercase=None , __lowercase=None , **__lowercase , ) -> List[Any]: super().__init__(**__lowercase ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {",".join(self.layer_types )}""" ) lowerCAmelCase_ : List[Any] = num_channels lowerCAmelCase_ : int = embedding_size lowerCAmelCase_ : str = hidden_sizes lowerCAmelCase_ : Union[str, Any] = depths lowerCAmelCase_ : Optional[Any] = layer_type lowerCAmelCase_ : Any = hidden_act lowerCAmelCase_ : Optional[Any] = downsample_in_first_stage lowerCAmelCase_ : int = ['''stem'''] + [f"""stage{idx}""" for idx in range(1 , len(__lowercase ) + 1 )] lowerCAmelCase_ , lowerCAmelCase_ : int = get_aligned_output_features_output_indices( out_features=__lowercase , out_indices=__lowercase , stage_names=self.stage_names ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = version.parse("""1.11""" ) @property def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowercase_ ( self ) -> float: return 1e-3
262
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _UpperCAmelCase : Union[str, Any] ="""pt""" elif is_tf_available(): _UpperCAmelCase : List[Any] ="""tf""" else: _UpperCAmelCase : Optional[int] ="""jax""" class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PerceiverTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = False def lowercase_ ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ : str = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase_ ( self ) -> Any: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def lowercase_ ( self , **__lowercase ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase=False , __lowercase=2_0 , __lowercase=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCAmelCase_ : Optional[Any] = [] for i in range(len(__lowercase ) ): try: lowerCAmelCase_ : List[str] = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowercase ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCAmelCase_ : List[str] = list(filter(lambda __lowercase : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , __lowercase ) ) lowerCAmelCase_ : Optional[int] = list(filter(lambda __lowercase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowercase ) , __lowercase ) ) if max_length is not None and len(__lowercase ) > max_length: lowerCAmelCase_ : Union[str, Any] = toks[:max_length] if min_length is not None and len(__lowercase ) < min_length and len(__lowercase ) > 0: while len(__lowercase ) < min_length: lowerCAmelCase_ : Union[str, Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCAmelCase_ : List[str] = [t[0] for t in toks] # Ensure consistency lowerCAmelCase_ : int = tokenizer.decode(__lowercase , clean_up_tokenization_spaces=__lowercase ) if " " not in output_txt and len(__lowercase ) > 1: lowerCAmelCase_ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowercase ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowercase ) ) if with_prefix_space: lowerCAmelCase_ : Any = ''' ''' + output_txt lowerCAmelCase_ : List[str] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) return output_txt, output_ids def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : List[str] = self.perceiver_tokenizer lowerCAmelCase_ : Any = '''Unicode €.''' lowerCAmelCase_ : Dict = tokenizer(__lowercase ) lowerCAmelCase_ : Any = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : str = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]Unicode €.[SEP]''' ) lowerCAmelCase_ : Optional[int] = tokenizer('''e è é ê ë''' ) lowerCAmelCase_ : str = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded['''input_ids'''] , __lowercase ) # decoding lowerCAmelCase_ : int = tokenizer.decode(__lowercase ) self.assertEqual(__lowercase , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Any = self.perceiver_tokenizer lowerCAmelCase_ : Dict = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowerCAmelCase_ : Optional[int] = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on lowerCAmelCase_ : Optional[int] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) self.assertIsInstance(__lowercase , __lowercase ) if FRAMEWORK != "jax": lowerCAmelCase_ : str = list(batch.input_ids.numpy()[0] ) else: lowerCAmelCase_ : Union[str, Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowercase , __lowercase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : int = self.perceiver_tokenizer lowerCAmelCase_ : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowerCAmelCase_ : List[Any] = tokenizer(__lowercase , padding=__lowercase , return_tensors=__lowercase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __lowercase ) self.assertIn('''attention_mask''' , __lowercase ) self.assertNotIn('''decoder_input_ids''' , __lowercase ) self.assertNotIn('''decoder_attention_mask''' , __lowercase ) def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.perceiver_tokenizer lowerCAmelCase_ : int = [ '''Summary of the text.''', '''Another summary.''', ] lowerCAmelCase_ : List[str] = tokenizer( text_target=__lowercase , max_length=3_2 , padding='''max_length''' , truncation=__lowercase , return_tensors=__lowercase ) self.assertEqual(3_2 , targets['''input_ids'''].shape[1] ) def lowercase_ ( self ) -> Optional[Any]: # safety check on max_len default value so we are sure the test works lowerCAmelCase_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test lowerCAmelCase_ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() lowerCAmelCase_ : str = ''' He is very happy, UNwant\u00E9d,running''' lowerCAmelCase_ : Optional[int] = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) shutil.rmtree(__lowercase ) lowerCAmelCase_ : Optional[int] = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowerCAmelCase_ : Optional[int] = tempfile.mkdtemp() lowerCAmelCase_ : List[str] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowerCAmelCase_ : Any = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowerCAmelCase_ : str = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) tokenizer.save_pretrained(__lowercase ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase ) lowerCAmelCase_ : Optional[Any] = after_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) lowerCAmelCase_ : str = tokenizer.__class__.from_pretrained(__lowercase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowercase ) def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Tuple = json.load(__lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowerCAmelCase_ : Any = json.load(__lowercase ) lowerCAmelCase_ : Optional[int] = [f"""<extra_id_{i}>""" for i in range(1_2_5 )] lowerCAmelCase_ : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowerCAmelCase_ : Any = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__lowercase , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) with open(os.path.join(__lowercase , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__lowercase , __lowercase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCAmelCase_ : int = tokenizer_class.from_pretrained( __lowercase , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCAmelCase_ : Tuple = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__lowercase )] lowerCAmelCase_ : Dict = tokenizer_class.from_pretrained( __lowercase , additional_special_tokens=__lowercase , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , '''�''' ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Any: pass def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> List[str]: pass def lowercase_ ( self ) -> Dict: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCAmelCase_ : Tuple = self.get_tokenizers(fast=__lowercase , do_lower_case=__lowercase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowerCAmelCase_ : List[str] = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowerCAmelCase_ : Optional[int] = tokenizer.convert_tokens_to_string(__lowercase ) self.assertIsInstance(__lowercase , __lowercase )
262
1
def lowerCAmelCase ( lowerCAmelCase_ )-> list: def merge(lowerCAmelCase_ , lowerCAmelCase_ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(lowerCAmelCase_ ) <= 1: return collection lowerCAmelCase_ : str = len(lowerCAmelCase_ ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() _UpperCAmelCase : int =input("""Enter numbers separated by a comma:\n""").strip() _UpperCAmelCase : Tuple =[int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
262
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case__: '''simple docstring''' @staticmethod def lowercase_ ( *__lowercase , **__lowercase ) -> Union[str, Any]: pass @is_pipeline_test @require_vision @require_torch class snake_case__( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : str = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def lowercase_ ( self , __lowercase , __lowercase ) -> str: lowerCAmelCase_ : Tuple = object_detector(examples[0] , threshold=0.0 ) lowerCAmelCase_ : Dict = len(__lowercase ) self.assertGreater(__lowercase , 0 ) self.assertEqual( __lowercase , [ { '''score''': ANY(__lowercase ), '''label''': ANY(__lowercase ), '''box''': {'''xmin''': ANY(__lowercase ), '''ymin''': ANY(__lowercase ), '''xmax''': ANY(__lowercase ), '''ymax''': ANY(__lowercase )}, } for i in range(__lowercase ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch def lowercase_ ( self ) -> int: lowerCAmelCase_ : Union[str, Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) lowerCAmelCase_ : Union[str, Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] , ) lowerCAmelCase_ : Union[str, Any] = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.72_35, '''label''': '''cat''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.72_18, '''label''': '''remote''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.71_84, '''label''': '''couch''', '''box''': {'''xmin''': 2_0_4, '''ymin''': 1_6_7, '''xmax''': 2_3_2, '''ymax''': 1_9_0}}, {'''score''': 0.67_48, '''label''': '''remote''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_56, '''label''': '''cat''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.66_14, '''label''': '''couch''', '''box''': {'''xmin''': 5_7_1, '''ymin''': 8_3, '''xmax''': 5_9_8, '''ymax''': 1_0_3}}, {'''score''': 0.64_56, '''label''': '''remote''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, {'''score''': 0.6_42, '''label''': '''remote''', '''box''': {'''xmin''': 6_7, '''ymin''': 2_7_4, '''xmax''': 9_3, '''ymax''': 2_9_7}}, {'''score''': 0.64_19, '''label''': '''cat''', '''box''': {'''xmin''': 4_9_4, '''ymin''': 1_0_5, '''xmax''': 5_2_1, '''ymax''': 1_2_7}}, ] ] , ) @require_torch @slow def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Any = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ] , ) lowerCAmelCase_ : Tuple = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, {'''score''': 0.14_74, '''label''': '''remote''', '''box''': {'''xmin''': 3_3_5, '''ymin''': 7_4, '''xmax''': 3_7_1, '''ymax''': 1_8_7}}, {'''score''': 0.12_08, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_4_2, '''ymax''': 4_7_6}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def lowercase_ ( self ) -> List[str]: pass @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Any = 0.2 lowerCAmelCase_ : List[Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, {'''score''': 0.25_37, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 5_5, '''xmax''': 3_1_5, '''ymax''': 4_7_2}}, ] , ) @require_torch @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : Dict = 2 lowerCAmelCase_ : Union[str, Any] = pipeline('''zero-shot-object-detection''' ) lowerCAmelCase_ : Optional[Any] = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=__lowercase , ) self.assertEqual( nested_simplify(__lowercase , decimals=4 ) , [ {'''score''': 0.28_68, '''label''': '''cat''', '''box''': {'''xmin''': 3_2_4, '''ymin''': 2_0, '''xmax''': 6_4_0, '''ymax''': 3_7_3}}, {'''score''': 0.2_77, '''label''': '''remote''', '''box''': {'''xmin''': 4_0, '''ymin''': 7_2, '''xmax''': 1_7_7, '''ymax''': 1_1_5}}, ] , )
262
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = """naver-clova-ix/donut-base-finetuned-docvqa""" SCREAMING_SNAKE_CASE__ : str = ( """This is a tool that answers a question about an document (pdf). It takes an input named `document` which """ """should be the document containing the information, as well as a `question` that is the question about the """ """document. It returns a text that contains the answer to the question.""" ) SCREAMING_SNAKE_CASE__ : str = """document_qa""" SCREAMING_SNAKE_CASE__ : Tuple = AutoProcessor SCREAMING_SNAKE_CASE__ : Optional[int] = VisionEncoderDecoderModel SCREAMING_SNAKE_CASE__ : List[str] = ["""image""", """text"""] SCREAMING_SNAKE_CASE__ : Any = ["""text"""] def __init__( self , *__lowercase , **__lowercase ) -> Any: if not is_vision_available(): raise ValueError('''Pillow must be installed to use the DocumentQuestionAnsweringTool.''' ) super().__init__(*__lowercase , **__lowercase ) def lowercase_ ( self , __lowercase , __lowercase ) -> Any: lowerCAmelCase_ : Union[str, Any] = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' lowerCAmelCase_ : List[Any] = task_prompt.replace('''{user_input}''' , __lowercase ) lowerCAmelCase_ : int = self.pre_processor.tokenizer( __lowercase , add_special_tokens=__lowercase , return_tensors='''pt''' ).input_ids lowerCAmelCase_ : str = self.pre_processor(__lowercase , return_tensors='''pt''' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def lowercase_ ( self , __lowercase ) -> int: return self.model.generate( inputs['''pixel_values'''].to(self.device ) , decoder_input_ids=inputs['''decoder_input_ids'''].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__lowercase , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__lowercase , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__lowercase , ).sequences def lowercase_ ( self , __lowercase ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = self.pre_processor.batch_decode(__lowercase )[0] lowerCAmelCase_ : List[str] = sequence.replace(self.pre_processor.tokenizer.eos_token , '''''' ) lowerCAmelCase_ : List[str] = sequence.replace(self.pre_processor.tokenizer.pad_token , '''''' ) lowerCAmelCase_ : List[str] = re.sub(R'''<.*?>''' , '''''' , __lowercase , count=1 ).strip() # remove first task start token lowerCAmelCase_ : Tuple = self.pre_processor.tokenajson(__lowercase ) return sequence["answer"]
262
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
1
def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> int: lowerCAmelCase_ : Any = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): lowerCAmelCase_ : Optional[Any] = n - k # Calculate C(n,k) for i in range(lowerCAmelCase_ ): result *= n - i result //= i + 1 return result def lowerCAmelCase ( lowerCAmelCase_ )-> int: return binomial_coefficient(2 * node_count , lowerCAmelCase_ ) // (node_count + 1) def lowerCAmelCase ( lowerCAmelCase_ )-> int: if n < 0: raise ValueError('''factorial() not defined for negative values''' ) lowerCAmelCase_ : List[str] = 1 for i in range(1 , n + 1 ): result *= i return result def lowerCAmelCase ( lowerCAmelCase_ )-> int: return catalan_number(lowerCAmelCase_ ) * factorial(lowerCAmelCase_ ) if __name__ == "__main__": _UpperCAmelCase : Dict =int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f"""Given {node_count} nodes, there are {binary_tree_count(node_count)} """ f"""binary trees and {catalan_number(node_count)} binary search trees.""" )
262
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _UpperCAmelCase : Dict ={ """susnato/ernie-m-base_pytorch""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json""", """susnato/ernie-m-large_pytorch""": """https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json""", } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ernie_m""" SCREAMING_SNAKE_CASE__ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowercase = 2_5_0_0_0_2 , __lowercase = 7_6_8 , __lowercase = 1_2 , __lowercase = 1_2 , __lowercase = 3_0_7_2 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 5_1_4 , __lowercase = 0.02 , __lowercase = 1 , __lowercase = 1e-05 , __lowercase=None , __lowercase=False , __lowercase=0.0 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , **__lowercase ) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : Tuple = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : int = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Any = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = classifier_dropout lowerCAmelCase_ : Any = is_decoder lowerCAmelCase_ : List[Any] = act_dropout
262
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _UpperCAmelCase : Dict ={"""configuration_encoder_decoder""": ["""EncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] =["""EncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] =["""TFEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Tuple =["""FlaxEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _UpperCAmelCase : str =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
262
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=1_8 , __lowercase=3_0 , __lowercase=4_0_0 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , ) -> List[Any]: lowerCAmelCase_ : Optional[Any] = size if size is not None else {'''shortest_edge''': 2_0} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : Tuple = image_size lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Dict = max_resolution lowerCAmelCase_ : Tuple = do_resize lowerCAmelCase_ : Optional[Any] = size lowerCAmelCase_ : Union[str, Any] = do_center_crop lowerCAmelCase_ : Optional[Any] = crop_size def lowercase_ ( self ) -> Union[str, Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = MobileNetVaImageProcessor if is_vision_available() else None def lowercase_ ( self ) -> List[str]: lowerCAmelCase_ : Union[str, Any] = MobileNetVaImageProcessingTester(self ) @property def lowercase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ ( self ) -> List[Any]: lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowercase , '''size''' ) ) self.assertTrue(hasattr(__lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowercase , '''crop_size''' ) ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size , {'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase_ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size , {'''height''': 8_4, '''width''': 8_4} ) def lowercase_ ( self ) -> Tuple: pass def lowercase_ ( self ) -> Union[str, Any]: # Initialize image_processing lowerCAmelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Optional[int]: # Initialize image_processing lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray ) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Tuple = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowercase_ ( self ) -> Any: # Initialize image_processing lowerCAmelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor ) # Test not batched input lowerCAmelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(__lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
262
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : Dict =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ={ """Helsinki-NLP/opus-mt-en-de""": """https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json""", # See all Marian models at https://huggingface.co/models?filter=marian } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = """marian""" SCREAMING_SNAKE_CASE__ : Any = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , __lowercase=5_8_1_0_1 , __lowercase=None , __lowercase=1_0_2_4 , __lowercase=1_2 , __lowercase=4_0_9_6 , __lowercase=1_6 , __lowercase=1_2 , __lowercase=4_0_9_6 , __lowercase=1_6 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=True , __lowercase=True , __lowercase="gelu" , __lowercase=1_0_2_4 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=5_8_1_0_0 , __lowercase=False , __lowercase=5_8_1_0_0 , __lowercase=0 , __lowercase=0 , __lowercase=True , **__lowercase , ) -> Dict: lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Union[str, Any] = decoder_vocab_size or vocab_size lowerCAmelCase_ : List[str] = max_position_embeddings lowerCAmelCase_ : Optional[int] = d_model lowerCAmelCase_ : List[Any] = encoder_ffn_dim lowerCAmelCase_ : str = encoder_layers lowerCAmelCase_ : Any = encoder_attention_heads lowerCAmelCase_ : Optional[int] = decoder_ffn_dim lowerCAmelCase_ : Optional[int] = decoder_layers lowerCAmelCase_ : Union[str, Any] = decoder_attention_heads lowerCAmelCase_ : Optional[int] = dropout lowerCAmelCase_ : str = attention_dropout lowerCAmelCase_ : List[str] = activation_dropout lowerCAmelCase_ : Any = activation_function lowerCAmelCase_ : str = init_std lowerCAmelCase_ : Optional[int] = encoder_layerdrop lowerCAmelCase_ : Optional[Any] = decoder_layerdrop lowerCAmelCase_ : Optional[int] = use_cache lowerCAmelCase_ : Optional[Any] = encoder_layers lowerCAmelCase_ : List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True lowerCAmelCase_ : Optional[int] = share_encoder_decoder_embeddings super().__init__( pad_token_id=__lowercase , eos_token_id=__lowercase , is_encoder_decoder=__lowercase , decoder_start_token_id=__lowercase , forced_eos_token_id=__lowercase , **__lowercase , ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : Union[str, Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowerCAmelCase_ : Any = {0: '''batch'''} lowerCAmelCase_ : Tuple = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowerCAmelCase_ : Dict = {0: '''batch''', 1: '''decoder_sequence'''} lowerCAmelCase_ : int = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCAmelCase_ : str = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowerCAmelCase_ , lowerCAmelCase_ : str = self.num_layers for i in range(__lowercase ): lowerCAmelCase_ : Any = {0: '''batch''', 2: '''past_sequence + sequence'''} lowerCAmelCase_ : List[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} else: lowerCAmelCase_ : Any = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : int = super().outputs else: lowerCAmelCase_ : Dict = super(__lowercase , self ).outputs if self.use_past: lowerCAmelCase_ , lowerCAmelCase_ : Tuple = self.num_layers for i in range(__lowercase ): lowerCAmelCase_ : Tuple = {0: '''batch''', 2: '''past_sequence + sequence'''} lowerCAmelCase_ : Optional[int] = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: lowerCAmelCase_ : Union[str, Any] = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Generate decoder inputs lowerCAmelCase_ : List[str] = seq_length if not self.use_past else 1 lowerCAmelCase_ : int = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : Any = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} lowerCAmelCase_ : Optional[Any] = dict(**__lowercase , **__lowercase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : int = common_inputs['''input_ids'''].shape lowerCAmelCase_ : Union[str, Any] = common_inputs['''decoder_input_ids'''].shape[1] lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self.num_attention_heads lowerCAmelCase_ : Union[str, Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCAmelCase_ : Union[str, Any] = decoder_seq_length + 3 lowerCAmelCase_ : Union[str, Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCAmelCase_ : str = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(__lowercase , __lowercase )] , dim=1 ) lowerCAmelCase_ : Optional[Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = self.num_layers lowerCAmelCase_ : List[str] = min(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[Any] = max(__lowercase , __lowercase ) - min_num_layers lowerCAmelCase_ : List[Any] = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(__lowercase ): common_inputs["past_key_values"].append( ( torch.zeros(__lowercase ), torch.zeros(__lowercase ), torch.zeros(__lowercase ), torch.zeros(__lowercase ), ) ) # TODO: test this. lowerCAmelCase_ : str = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(__lowercase , __lowercase ): common_inputs["past_key_values"].append((torch.zeros(__lowercase ), torch.zeros(__lowercase )) ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: lowerCAmelCase_ : Optional[int] = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : Dict = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowerCAmelCase_ : Dict = seqlen + 2 lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self.num_layers lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self.num_attention_heads lowerCAmelCase_ : List[Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCAmelCase_ : Any = common_inputs['''attention_mask'''].dtype lowerCAmelCase_ : Union[str, Any] = torch.cat( [common_inputs['''attention_mask'''], torch.ones(__lowercase , __lowercase , dtype=__lowercase )] , dim=1 ) lowerCAmelCase_ : Tuple = [ (torch.zeros(__lowercase ), torch.zeros(__lowercase )) for _ in range(__lowercase ) ] return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowerCAmelCase_ : Dict = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCAmelCase_ : Optional[Any] = tokenizer.num_special_tokens_to_add(__lowercase ) lowerCAmelCase_ : Dict = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowercase ) # Generate dummy inputs according to compute batch and sequence lowerCAmelCase_ : List[str] = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCAmelCase_ : Optional[Any] = dict(tokenizer(__lowercase , return_tensors=__lowercase ) ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : Optional[int] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase ) else: lowerCAmelCase_ : Optional[int] = self._generate_dummy_inputs_for_causal_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase ) return common_inputs def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> List[str]: if self.task in ["default", "seq2seq-lm"]: lowerCAmelCase_ : List[str] = super()._flatten_past_key_values_(__lowercase , __lowercase , __lowercase , __lowercase ) else: lowerCAmelCase_ : Optional[Any] = super(__lowercase , self )._flatten_past_key_values_( __lowercase , __lowercase , __lowercase , __lowercase ) @property def lowercase_ ( self ) -> float: return 1e-4
262
from __future__ import annotations import math class snake_case__: '''simple docstring''' def __init__( self , __lowercase ) -> None: lowerCAmelCase_ : str = size # approximate the overall size of segment tree with given value lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] # create array to store lazy update lowerCAmelCase_ : Dict = [0 for i in range(0 , 4 * size )] lowerCAmelCase_ : Optional[int] = [0 for i in range(0 , 4 * size )] # flag for lazy update def lowercase_ ( self , __lowercase ) -> int: return idx * 2 def lowercase_ ( self , __lowercase ) -> int: return idx * 2 + 1 def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase ) -> None: if left_element == right_element: lowerCAmelCase_ : Tuple = a[left_element - 1] else: lowerCAmelCase_ : int = (left_element + right_element) // 2 self.build(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase ) self.build(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase ) lowerCAmelCase_ : Any = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> bool: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Union[str, Any] = False if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Any = self.lazy[idx] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: lowerCAmelCase_ : Dict = val if left_element != right_element: lowerCAmelCase_ : Union[str, Any] = val lowerCAmelCase_ : List[Any] = val lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : List[str] = True return True lowerCAmelCase_ : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) self.update(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : int = max( self.segment_tree[self.left(__lowercase )] , self.segment_tree[self.right(__lowercase )] ) return True def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> int | float: if self.flag[idx] is True: lowerCAmelCase_ : Union[str, Any] = self.lazy[idx] lowerCAmelCase_ : Optional[Any] = False if left_element != right_element: lowerCAmelCase_ : List[Any] = self.lazy[idx] lowerCAmelCase_ : Dict = self.lazy[idx] lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] lowerCAmelCase_ : List[Any] = (left_element + right_element) // 2 lowerCAmelCase_ : Tuple = self.query(self.left(__lowercase ) , __lowercase , __lowercase , __lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = self.query(self.right(__lowercase ) , mid + 1 , __lowercase , __lowercase , __lowercase ) return max(__lowercase , __lowercase ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , __lowercase , __lowercase ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _UpperCAmelCase : str =[1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _UpperCAmelCase : List[str] =15 _UpperCAmelCase : Any =SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
262
1
from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = DistilBertTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = DistilBertTokenizerFast SCREAMING_SNAKE_CASE__ : Any = True @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : int = DistilBertTokenizer.from_pretrained('''distilbert-base-uncased''' ) lowerCAmelCase_ : Any = tokenizer.encode('''sequence builders''' , add_special_tokens=__lowercase ) lowerCAmelCase_ : int = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__lowercase ) lowerCAmelCase_ : Optional[int] = tokenizer.build_inputs_with_special_tokens(__lowercase ) lowerCAmelCase_ : List[str] = tokenizer.build_inputs_with_special_tokens(__lowercase , __lowercase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
262
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py _UpperCAmelCase : Optional[int] ="""src/transformers""" _UpperCAmelCase : str ="""docs/source/en""" _UpperCAmelCase : Optional[int] =""".""" def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> List[str]: with open(lowerCAmelCase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase_ : int = f.readlines() # Find the start prompt. lowerCAmelCase_ : List[Any] = 0 while not lines[start_index].startswith(lowerCAmelCase_ ): start_index += 1 start_index += 1 lowerCAmelCase_ : List[str] = start_index while not lines[end_index].startswith(lowerCAmelCase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | _UpperCAmelCase : Optional[Any] ="""Model|Encoder|Decoder|ForConditionalGeneration""" # Regexes that match TF/Flax/PT model names. _UpperCAmelCase : Optional[int] =re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _UpperCAmelCase : Dict =re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _UpperCAmelCase : Optional[Any] =re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : Optional[int] =direct_transformers_import(TRANSFORMERS_PATH) def lowerCAmelCase ( lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : str = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' , lowerCAmelCase_ ) return [m.group(0 ) for m in matches] def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ )-> Optional[int]: lowerCAmelCase_ : Tuple = 2 if text == '''✅''' or text == '''❌''' else len(lowerCAmelCase_ ) lowerCAmelCase_ : int = (width - text_length) // 2 lowerCAmelCase_ : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def lowerCAmelCase ( )-> str: lowerCAmelCase_ : Any = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowerCAmelCase_ : Dict = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } lowerCAmelCase_ : List[Any] = {name: config.replace('''Config''' , '''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. lowerCAmelCase_ : Tuple = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[Any] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : Optional[int] = collections.defaultdict(lowerCAmelCase_ ) lowerCAmelCase_ : List[str] = collections.defaultdict(lowerCAmelCase_ ) # Let's lookup through all transformers object (once). for attr_name in dir(lowerCAmelCase_ ): lowerCAmelCase_ : Optional[int] = None if attr_name.endswith('''Tokenizer''' ): lowerCAmelCase_ : Union[str, Any] = slow_tokenizers lowerCAmelCase_ : List[str] = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): lowerCAmelCase_ : int = fast_tokenizers lowerCAmelCase_ : Union[str, Any] = attr_name[:-13] elif _re_tf_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = tf_models lowerCAmelCase_ : str = _re_tf_models.match(lowerCAmelCase_ ).groups()[0] elif _re_flax_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Tuple = flax_models lowerCAmelCase_ : Union[str, Any] = _re_flax_models.match(lowerCAmelCase_ ).groups()[0] elif _re_pt_models.match(lowerCAmelCase_ ) is not None: lowerCAmelCase_ : Any = pt_models lowerCAmelCase_ : List[Any] = _re_pt_models.match(lowerCAmelCase_ ).groups()[0] if lookup_dict is not None: while len(lowerCAmelCase_ ) > 0: if attr_name in model_name_to_prefix.values(): lowerCAmelCase_ : Union[str, Any] = True break # Try again after removing the last word in the name lowerCAmelCase_ : Any = ''''''.join(camel_case_split(lowerCAmelCase_ )[:-1] ) # Let's build that table! lowerCAmelCase_ : int = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) lowerCAmelCase_ : Tuple = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). lowerCAmelCase_ : Union[str, Any] = [len(lowerCAmelCase_ ) + 2 for c in columns] lowerCAmelCase_ : Optional[Any] = max([len(lowerCAmelCase_ ) for name in model_names] ) + 2 # Build the table per se lowerCAmelCase_ : Dict = '''|''' + '''|'''.join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for c, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" lowerCAmelCase_ : List[str] = {True: '''✅''', False: '''❌'''} for name in model_names: lowerCAmelCase_ : List[Any] = model_name_to_prefix[name] lowerCAmelCase_ : Union[str, Any] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(lowerCAmelCase_ , lowerCAmelCase_ ) for l, w in zip(lowerCAmelCase_ , lowerCAmelCase_ )] ) + "|\n" return table def lowerCAmelCase ( lowerCAmelCase_=False )-> Tuple: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = _find_text_in_file( filename=os.path.join(lowerCAmelCase_ , '''index.md''' ) , start_prompt='''<!--This table is updated automatically from the auto modules''' , end_prompt='''<!-- End table-->''' , ) lowerCAmelCase_ : Tuple = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(lowerCAmelCase_ , '''index.md''' ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase : Tuple =parser.parse_args() check_model_table(args.fix_and_overwrite)
262
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
262
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
1
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def lowerCAmelCase ( lowerCAmelCase_ )-> List[str]: if isinstance(lowerCAmelCase_ , collections.abc.Iterable ): return x return (x, x) @require_flax class snake_case__: '''simple docstring''' def lowercase_ ( self , __lowercase , __lowercase ) -> str: pass def lowercase_ ( self ) -> Dict: pass def lowercase_ ( self ) -> List[Any]: pass def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Optional[Any] = np.abs((a - b) ).max() self.assertLessEqual(__lowercase , __lowercase , f"""Difference between torch and flax is {diff} (>= {tol}).""" ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None , **__lowercase ) -> int: lowerCAmelCase_ : Any = VisionTextDualEncoderConfig.from_vision_text_configs(__lowercase , __lowercase ) lowerCAmelCase_ : Union[str, Any] = FlaxVisionTextDualEncoderModel(__lowercase ) lowerCAmelCase_ : List[Any] = model(input_ids=__lowercase , pixel_values=__lowercase , attention_mask=__lowercase ) self.assertEqual(output['''text_embeds'''].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape , (pixel_values.shape[0], config.projection_dim) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None , **__lowercase ) -> List[Any]: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self.get_vision_text_model(__lowercase , __lowercase ) lowerCAmelCase_ : List[Any] = {'''vision_model''': vision_model, '''text_model''': text_model} lowerCAmelCase_ : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__lowercase ) lowerCAmelCase_ : Tuple = model(input_ids=__lowercase , pixel_values=__lowercase , attention_mask=__lowercase ) self.assertEqual(output['''text_embeds'''].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None , **__lowercase ) -> Any: lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.get_vision_text_model(__lowercase , __lowercase ) lowerCAmelCase_ : str = {'''vision_model''': vision_model, '''text_model''': text_model} lowerCAmelCase_ : Tuple = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__lowercase ) lowerCAmelCase_ : Union[str, Any] = model(input_ids=__lowercase , pixel_values=__lowercase , attention_mask=__lowercase ) lowerCAmelCase_ : Optional[Any] = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowercase ) lowerCAmelCase_ : int = FlaxVisionTextDualEncoderModel.from_pretrained(__lowercase ) lowerCAmelCase_ : Any = model(input_ids=__lowercase , pixel_values=__lowercase , attention_mask=__lowercase ) lowerCAmelCase_ : List[str] = after_output[0] lowerCAmelCase_ : Union[str, Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__lowercase , 1e-3 ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None , **__lowercase ) -> List[Any]: lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.get_vision_text_model(__lowercase , __lowercase ) lowerCAmelCase_ : List[str] = {'''vision_model''': vision_model, '''text_model''': text_model} lowerCAmelCase_ : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__lowercase ) lowerCAmelCase_ : Union[str, Any] = model( input_ids=__lowercase , pixel_values=__lowercase , attention_mask=__lowercase , output_attentions=__lowercase ) lowerCAmelCase_ : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(__lowercase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase_ : Union[str, Any] = to_atuple(vision_model.config.image_size ) lowerCAmelCase_ : int = to_atuple(vision_model.config.patch_size ) lowerCAmelCase_ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowerCAmelCase_ : int = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) lowerCAmelCase_ : int = output.text_model_output.attentions self.assertEqual(len(__lowercase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: pt_model.to(__lowercase ) pt_model.eval() # prepare inputs lowerCAmelCase_ : Dict = inputs_dict lowerCAmelCase_ : Optional[int] = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): lowerCAmelCase_ : Dict = pt_model(**__lowercase ).to_tuple() lowerCAmelCase_ : str = fx_model(**__lowercase ).to_tuple() self.assertEqual(len(__lowercase ) , len(__lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(__lowercase , pt_output.numpy() , 4e-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(__lowercase ) lowerCAmelCase_ : Tuple = FlaxVisionTextDualEncoderModel.from_pretrained(__lowercase , from_pt=__lowercase ) lowerCAmelCase_ : Optional[Any] = fx_model_loaded(**__lowercase ).to_tuple() self.assertEqual(len(__lowercase ) , len(__lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(__lowercase , pt_output.numpy() , 4e-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(__lowercase ) lowerCAmelCase_ : Optional[int] = VisionTextDualEncoderModel.from_pretrained(__lowercase , from_flax=__lowercase ) pt_model_loaded.to(__lowercase ) pt_model_loaded.eval() with torch.no_grad(): lowerCAmelCase_ : List[Any] = pt_model_loaded(**__lowercase ).to_tuple() self.assertEqual(len(__lowercase ) , len(__lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(__lowercase , pt_output_loaded.numpy() , 4e-2 ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(__lowercase , __lowercase ) lowerCAmelCase_ : Tuple = VisionTextDualEncoderModel(__lowercase ) lowerCAmelCase_ : Union[str, Any] = FlaxVisionTextDualEncoderModel(__lowercase ) lowerCAmelCase_ : Optional[int] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , __lowercase ) lowerCAmelCase_ : Dict = fx_state self.check_pt_flax_equivalence(__lowercase , __lowercase , __lowercase ) def lowercase_ ( self , __lowercase , __lowercase , __lowercase ) -> List[str]: lowerCAmelCase_ : Any = VisionTextDualEncoderConfig.from_vision_text_configs(__lowercase , __lowercase ) lowerCAmelCase_ : int = VisionTextDualEncoderModel(__lowercase ) lowerCAmelCase_ : List[Any] = FlaxVisionTextDualEncoderModel(__lowercase ) lowerCAmelCase_ : Dict = load_flax_weights_in_pytorch_model(__lowercase , fx_model.params ) self.check_pt_flax_equivalence(__lowercase , __lowercase , __lowercase ) def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : Any = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**__lowercase ) def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ : str = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**__lowercase ) def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : int = self.prepare_config_and_inputs() self.check_save_load(**__lowercase ) def lowercase_ ( self ) -> Dict: lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**__lowercase ) @is_pt_flax_cross_test def lowercase_ ( self ) -> Union[str, Any]: lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() lowerCAmelCase_ : Tuple = config_inputs_dict.pop('''vision_config''' ) lowerCAmelCase_ : Optional[Any] = config_inputs_dict.pop('''text_config''' ) lowerCAmelCase_ : List[Any] = config_inputs_dict self.check_equivalence_pt_to_flax(__lowercase , __lowercase , __lowercase ) self.check_equivalence_flax_to_pt(__lowercase , __lowercase , __lowercase ) @slow def lowercase_ ( self ) -> Optional[int]: lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.get_pretrained_model_and_inputs() lowerCAmelCase_ : Tuple = model_a(**__lowercase ) lowerCAmelCase_ : int = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(__lowercase ) lowerCAmelCase_ : Any = FlaxVisionTextDualEncoderModel.from_pretrained(__lowercase ) lowerCAmelCase_ : str = model_a(**__lowercase ) lowerCAmelCase_ : int = after_outputs[0] lowerCAmelCase_ : Any = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__lowercase , 1e-5 ) @require_flax class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-vit''' , '''hf-internal-testing/tiny-bert''' , vision_from_pt=__lowercase , text_from_pt=__lowercase , ) lowerCAmelCase_ : int = 1_3 lowerCAmelCase_ : Any = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCAmelCase_ : str = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) lowerCAmelCase_ : List[str] = random_attention_mask([batch_size, 4] ) lowerCAmelCase_ : List[str] = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def lowercase_ ( self , __lowercase , __lowercase ) -> Dict: lowerCAmelCase_ : Any = FlaxViTModel(__lowercase ) lowerCAmelCase_ : List[Any] = FlaxBertModel(__lowercase ) return vision_model, text_model def lowercase_ ( self ) -> Optional[Any]: lowerCAmelCase_ : List[str] = FlaxViTModelTester(self ) lowerCAmelCase_ : List[str] = FlaxBertModelTester(self ) lowerCAmelCase_ : Dict = vit_model_tester.prepare_config_and_inputs() lowerCAmelCase_ : int = bert_model_tester.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ : Tuple = vision_config_and_inputs lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class snake_case__( UpperCAmelCase__, unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> Any: lowerCAmelCase_ : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-clip''' , '''hf-internal-testing/tiny-bert''' , vision_from_pt=__lowercase , text_from_pt=__lowercase , ) lowerCAmelCase_ : Optional[int] = 1_3 lowerCAmelCase_ : Union[str, Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCAmelCase_ : Tuple = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) lowerCAmelCase_ : List[Any] = random_attention_mask([batch_size, 4] ) lowerCAmelCase_ : List[Any] = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def lowercase_ ( self , __lowercase , __lowercase ) -> Optional[int]: lowerCAmelCase_ : Optional[int] = FlaxCLIPVisionModel(__lowercase ) lowerCAmelCase_ : Tuple = FlaxBertModel(__lowercase ) return vision_model, text_model def lowercase_ ( self ) -> str: lowerCAmelCase_ : Any = FlaxCLIPVisionModelTester(self ) lowerCAmelCase_ : List[Any] = FlaxBertModelTester(self ) lowerCAmelCase_ : Optional[int] = clip_model_tester.prepare_config_and_inputs() lowerCAmelCase_ : Optional[int] = bert_model_tester.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = vision_config_and_inputs lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class snake_case__( unittest.TestCase ): '''simple docstring''' @slow def lowercase_ ( self ) -> str: lowerCAmelCase_ : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained('''clip-italian/clip-italian''' , logit_scale_init_value=1.0 ) lowerCAmelCase_ : Optional[Any] = VisionTextDualEncoderProcessor.from_pretrained('''clip-italian/clip-italian''' ) lowerCAmelCase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowerCAmelCase_ : Optional[Any] = processor( text=['''una foto di un gatto''', '''una foto di un cane'''] , images=__lowercase , padding=__lowercase , return_tensors='''np''' ) lowerCAmelCase_ : List[str] = model(**__lowercase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) lowerCAmelCase_ : Optional[Any] = np.array([[1.2_28_47_27, 0.3_10_41_22]] ) self.assertTrue(np.allclose(outputs.logits_per_image , __lowercase , atol=1e-3 ) )
262
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : Tuple =None _UpperCAmelCase : int =logging.get_logger(__name__) _UpperCAmelCase : Dict ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Any ={ """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int ={ """facebook/nllb-large-en-ro""": 1024, """facebook/nllb-200-distilled-600M""": 1024, } # fmt: off _UpperCAmelCase : Any =["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : int = NllbTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=False , **__lowercase , ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : int = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token lowerCAmelCase_ : List[Any] = legacy_behaviour super().__init__( vocab_file=__lowercase , tokenizer_file=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , additional_special_tokens=__lowercase , legacy_behaviour=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_file lowerCAmelCase_ : List[Any] = False if not self.vocab_file else True lowerCAmelCase_ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) lowerCAmelCase_ : Optional[Any] = { lang_code: self.convert_tokens_to_ids(__lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase_ : Any = src_lang if src_lang is not None else '''eng_Latn''' lowerCAmelCase_ : str = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase_ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowercase_ ( self ) -> str: return self._src_lang @src_lang.setter def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ ( self , __lowercase , __lowercase = None ) -> List[int]: lowerCAmelCase_ : Optional[Any] = [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase ) -> str: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : int = self(__lowercase , add_special_tokens=__lowercase , return_tensors=__lowercase , **__lowercase ) lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) lowerCAmelCase_ : List[Any] = tgt_lang_id return inputs def lowercase_ ( self , __lowercase , __lowercase = "eng_Latn" , __lowercase = None , __lowercase = "fra_Latn" , **__lowercase , ) -> BatchEncoding: lowerCAmelCase_ : List[str] = src_lang lowerCAmelCase_ : List[str] = tgt_lang return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def lowercase_ ( self ) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : List[str] = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[str] = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Optional[int] = [self.cur_lang_code] lowerCAmelCase_ : List[Any] = [self.eos_token_id] lowerCAmelCase_ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Any = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase ) -> None: lowerCAmelCase_ : Dict = self.convert_tokens_to_ids(__lowercase ) if self.legacy_behaviour: lowerCAmelCase_ : List[Any] = [] lowerCAmelCase_ : Any = [self.eos_token_id, self.cur_lang_code] else: lowerCAmelCase_ : Any = [self.cur_lang_code] lowerCAmelCase_ : Any = [self.eos_token_id] lowerCAmelCase_ : List[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase_ : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowercase_ ( self , __lowercase , __lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__lowercase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCAmelCase_ : Any = os.path.join( __lowercase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowercase ): copyfile(self.vocab_file , __lowercase ) return (out_vocab_file,)
262
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase__ = logging.get_logger(__name__) class lowercase_ ( lowercase , lowercase ): '''simple docstring''' __snake_case = '''maskformer-swin''' __snake_case = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , __UpperCAmelCase : Tuple=224 , __UpperCAmelCase : Tuple=4 , __UpperCAmelCase : int=3 , __UpperCAmelCase : Optional[Any]=96 , __UpperCAmelCase : Tuple=[2, 2, 6, 2] , __UpperCAmelCase : Optional[Any]=[3, 6, 12, 24] , __UpperCAmelCase : Dict=7 , __UpperCAmelCase : Tuple=4.0 , __UpperCAmelCase : Tuple=True , __UpperCAmelCase : Tuple=0.0 , __UpperCAmelCase : Optional[Any]=0.0 , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : str="gelu" , __UpperCAmelCase : Tuple=False , __UpperCAmelCase : List[str]=0.02 , __UpperCAmelCase : List[Any]=1e-5 , __UpperCAmelCase : int=None , __UpperCAmelCase : str=None , **__UpperCAmelCase : str , ) ->List[str]: """simple docstring""" super().__init__(**__UpperCAmelCase ) a = image_size a = patch_size a = num_channels a = embed_dim a = depths a = len(__UpperCAmelCase ) a = num_heads a = window_size a = mlp_ratio a = qkv_bias a = hidden_dropout_prob a = attention_probs_dropout_prob a = drop_path_rate a = hidden_act a = use_absolute_embeddings a = layer_norm_eps a = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model a = int(embed_dim * 2 ** (len(__UpperCAmelCase ) - 1) ) a = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(__UpperCAmelCase ) + 1 )] a , a = get_aligned_output_features_output_indices( out_features=__UpperCAmelCase , out_indices=__UpperCAmelCase , stage_names=self.stage_names )
0
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _UpperCAmelCase : Optional[Any] =NewType("""DataClass""", Any) _UpperCAmelCase : Dict =NewType("""DataClassType""", Any) def lowerCAmelCase ( lowerCAmelCase_ )-> Tuple: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase ( lowerCAmelCase_ )-> Callable[[str], Any]: lowerCAmelCase_ : str = {str(lowerCAmelCase_ ): choice for choice in choices} return lambda lowerCAmelCase_ : str_to_choice.get(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCAmelCase ( *, lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = dataclasses.MISSING , lowerCAmelCase_ = None , **lowerCAmelCase_ , )-> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase_ : Dict = {} if aliases is not None: lowerCAmelCase_ : str = aliases if help is not None: lowerCAmelCase_ : Tuple = help return dataclasses.field(metadata=lowerCAmelCase_ , default=lowerCAmelCase_ , default_factory=lowerCAmelCase_ , **lowerCAmelCase_ ) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Iterable[DataClassType] def __init__( self , __lowercase , **__lowercase ) -> List[str]: # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase_ : Optional[int] = ArgumentDefaultsHelpFormatter super().__init__(**__lowercase ) if dataclasses.is_dataclass(__lowercase ): lowerCAmelCase_ : Union[str, Any] = [dataclass_types] lowerCAmelCase_ : List[Any] = list(__lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowercase ) @staticmethod def lowercase_ ( __lowercase , __lowercase ) -> Union[str, Any]: lowerCAmelCase_ : Optional[Any] = f"""--{field.name}""" lowerCAmelCase_ : Tuple = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowercase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase_ : List[str] = kwargs.pop('''aliases''' , [] ) if isinstance(__lowercase , __lowercase ): lowerCAmelCase_ : Optional[Any] = [aliases] lowerCAmelCase_ : Any = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(__lowercase , '''UnionType''' ) and isinstance(__lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowercase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(__lowercase ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase_ : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase_ : Dict = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase_ : str = ( field.type.__args__[0] if isinstance(__lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase_ : List[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase_ : Dict = {} if origin_type is Literal or (isinstance(field.type , __lowercase ) and issubclass(field.type , __lowercase )): if origin_type is Literal: lowerCAmelCase_ : Optional[Any] = field.type.__args__ else: lowerCAmelCase_ : int = [x.value for x in field.type] lowerCAmelCase_ : str = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase_ : str = field.default else: lowerCAmelCase_ : Tuple = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase_ : Tuple = copy(__lowercase ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase_ : Dict = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase_ : List[str] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase_ : int = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase_ : List[Any] = True elif isclass(__lowercase ) and issubclass(__lowercase , __lowercase ): lowerCAmelCase_ : Union[str, Any] = field.type.__args__[0] lowerCAmelCase_ : Dict = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : Any = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase_ : Dict = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase_ : List[Any] = field.default_factory() else: lowerCAmelCase_ : int = True parser.add_argument(__lowercase , *__lowercase , **__lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase_ : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **__lowercase ) def lowercase_ ( self , __lowercase ) -> List[Any]: if hasattr(__lowercase , '''_argument_group_name''' ): lowerCAmelCase_ : str = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase_ : Dict = self try: lowerCAmelCase_ : Dict[str, type] = get_type_hints(__lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(__lowercase ): lowerCAmelCase_ : Any = '''.'''.join(map(__lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(__lowercase ): if not field.init: continue lowerCAmelCase_ : Optional[int] = type_hints[field.name] self._parse_dataclass_field(__lowercase , __lowercase ) def lowercase_ ( self , __lowercase=None , __lowercase=False , __lowercase=True , __lowercase=None , __lowercase=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase_ : str = [] if args_filename: args_files.append(Path(__lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase_ : str = ArgumentParser() args_file_parser.add_argument(__lowercase , type=__lowercase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = args_file_parser.parse_known_args(args=__lowercase ) lowerCAmelCase_ : int = vars(__lowercase ).get(args_file_flag.lstrip('''-''' ) , __lowercase ) if cmd_args_file_paths: args_files.extend([Path(__lowercase ) for p in cmd_args_file_paths] ) lowerCAmelCase_ : Dict = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase_ , lowerCAmelCase_ : Dict = self.parse_known_args(args=__lowercase ) lowerCAmelCase_ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase_ : str = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : str = {k: v for k, v in vars(__lowercase ).items() if k in keys} for k in keys: delattr(__lowercase , __lowercase ) lowerCAmelCase_ : Optional[int] = dtype(**__lowercase ) outputs.append(__lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : int = set(args.keys() ) lowerCAmelCase_ : str = [] for dtype in self.dataclass_types: lowerCAmelCase_ : int = {f.name for f in dataclasses.fields(__lowercase ) if f.init} lowerCAmelCase_ : List[str] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase_ : List[str] = dtype(**__lowercase ) outputs.append(__lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(__lowercase )}""" ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: with open(Path(__lowercase ) , encoding='''utf-8''' ) as open_json_file: lowerCAmelCase_ : Dict = json.loads(open_json_file.read() ) lowerCAmelCase_ : str = self.parse_dict(__lowercase , allow_extra_keys=__lowercase ) return tuple(__lowercase ) def lowercase_ ( self , __lowercase , __lowercase = False ) -> Tuple[DataClass, ...]: lowerCAmelCase_ : Optional[Any] = self.parse_dict(yaml.safe_load(Path(__lowercase ).read_text() ) , allow_extra_keys=__lowercase ) return tuple(__lowercase )
262
0