code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase__ = { '''configuration_blenderbot_small''': [ '''BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotSmallConfig''', '''BlenderbotSmallOnnxConfig''', ], '''tokenization_blenderbot_small''': ['''BlenderbotSmallTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['''BlenderbotSmallTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotSmallForCausalLM''', '''BlenderbotSmallForConditionalGeneration''', '''BlenderbotSmallModel''', '''BlenderbotSmallPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TFBlenderbotSmallForConditionalGeneration''', '''TFBlenderbotSmallModel''', '''TFBlenderbotSmallPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''FlaxBlenderbotSmallForConditionalGeneration''', '''FlaxBlenderbotSmallModel''', '''FlaxBlenderbotSmallPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
from abc import ABC, abstractmethod from typing import List, Optional class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[str]) -> Any: """simple docstring""" self.test() def __UpperCAmelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = False while not completed: if counter == 1: self.reset() _UpperCamelCase = self.advance() if not self.does_advance(__UpperCamelCase): raise Exception( "Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.update(__UpperCamelCase) counter += 1 if counter > 10000: raise Exception("update() does not fulfill the constraint.") if self.remaining() != 0: raise Exception("Custom Constraint is not defined correctly.") @abstractmethod def __UpperCAmelCase ( self : str) -> Any: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') @abstractmethod def __UpperCAmelCase ( self : Optional[int] , lowercase_ : int) -> Dict: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') @abstractmethod def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int) -> Union[str, Any]: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') @abstractmethod def __UpperCAmelCase ( self : List[str]) -> Union[str, Any]: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') @abstractmethod def __UpperCAmelCase ( self : List[str]) -> str: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') @abstractmethod def __UpperCAmelCase ( self : str , lowercase_ : List[str]=False) -> Any: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Tuple , lowercase_ : List[int]) -> Optional[Any]: """simple docstring""" super(__UpperCamelCase , self).__init__() if not isinstance(__UpperCamelCase , __UpperCamelCase) or len(__UpperCamelCase) == 0: raise ValueError(f'`token_ids` has to be a non-empty list, but is {token_ids}.') if any((not isinstance(__UpperCamelCase , __UpperCamelCase) or token_id < 0) for token_id in token_ids): raise ValueError(f'Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.') _UpperCamelCase = token_ids _UpperCamelCase = len(self.token_ids) _UpperCamelCase = -1 # the index of the currently fulfilled step _UpperCamelCase = False def __UpperCAmelCase ( self : Optional[int]) -> Dict: """simple docstring""" if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def __UpperCAmelCase ( self : List[Any] , lowercase_ : int) -> Dict: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(__UpperCamelCase)}') if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def __UpperCAmelCase ( self : int , lowercase_ : int) -> List[str]: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(__UpperCamelCase)}') _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False if self.does_advance(__UpperCamelCase): self.fulfilled_idx += 1 _UpperCamelCase = True if self.fulfilled_idx == (self.seqlen - 1): _UpperCamelCase = True _UpperCamelCase = completed else: # failed to make progress. _UpperCamelCase = True self.reset() return stepped, completed, reset def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" _UpperCamelCase = False _UpperCamelCase = 0 def __UpperCAmelCase ( self : Optional[Any]) -> Dict: """simple docstring""" return self.seqlen - (self.fulfilled_idx + 1) def __UpperCAmelCase ( self : int , lowercase_ : Union[str, Any]=False) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = PhrasalConstraint(self.token_ids) if stateful: _UpperCamelCase = self.seqlen _UpperCamelCase = self.fulfilled_idx _UpperCamelCase = self.completed return new_constraint class _UpperCAmelCase : '''simple docstring''' def __init__( self : int , lowercase_ : List[List[int]] , lowercase_ : str=True) -> Dict: """simple docstring""" _UpperCamelCase = max([len(__UpperCamelCase) for one in nested_token_ids]) _UpperCamelCase = {} for token_ids in nested_token_ids: _UpperCamelCase = root for tidx, token_id in enumerate(__UpperCamelCase): if token_id not in level: _UpperCamelCase = {} _UpperCamelCase = level[token_id] if no_subsets and self.has_subsets(__UpperCamelCase , __UpperCamelCase): raise ValueError( "Each list in `nested_token_ids` can't be a complete subset of another list, but is" f' {nested_token_ids}.') _UpperCamelCase = root def __UpperCAmelCase ( self : List[str] , lowercase_ : Union[str, Any]) -> Dict: """simple docstring""" _UpperCamelCase = self.trie for current_token in current_seq: _UpperCamelCase = start[current_token] _UpperCamelCase = list(start.keys()) return next_tokens def __UpperCAmelCase ( self : int , lowercase_ : List[str]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.next_tokens(__UpperCamelCase) return len(__UpperCamelCase) == 0 def __UpperCAmelCase ( self : List[Any] , lowercase_ : int) -> Optional[int]: """simple docstring""" _UpperCamelCase = list(root.values()) if len(__UpperCamelCase) == 0: return 1 else: return sum([self.count_leaves(__UpperCamelCase) for nn in next_nodes]) def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Any) -> str: """simple docstring""" _UpperCamelCase = self.count_leaves(__UpperCamelCase) return len(__UpperCamelCase) != leaf_count class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : List[List[int]]) -> str: """simple docstring""" super(__UpperCamelCase , self).__init__() if not isinstance(__UpperCamelCase , __UpperCamelCase) or len(__UpperCamelCase) == 0: raise ValueError(f'`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.') if any(not isinstance(__UpperCamelCase , __UpperCamelCase) for token_ids in nested_token_ids): raise ValueError(f'`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.') if any( any((not isinstance(__UpperCamelCase , __UpperCamelCase) or token_id < 0) for token_id in token_ids) for token_ids in nested_token_ids): raise ValueError( f'Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.') _UpperCamelCase = DisjunctiveTrie(__UpperCamelCase) _UpperCamelCase = nested_token_ids _UpperCamelCase = self.trie.max_height _UpperCamelCase = [] _UpperCamelCase = False def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = self.trie.next_tokens(self.current_seq) if len(__UpperCamelCase) == 0: return None else: return token_list def __UpperCAmelCase ( self : Any , lowercase_ : int) -> List[Any]: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(__UpperCamelCase)}') _UpperCamelCase = self.trie.next_tokens(self.current_seq) return token_id in next_tokens def __UpperCAmelCase ( self : Optional[int] , lowercase_ : int) -> Tuple: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(__UpperCamelCase)}') _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False if self.does_advance(__UpperCamelCase): self.current_seq.append(__UpperCamelCase) _UpperCamelCase = True else: _UpperCamelCase = True self.reset() _UpperCamelCase = self.trie.reached_leaf(self.current_seq) _UpperCamelCase = completed return stepped, completed, reset def __UpperCAmelCase ( self : str) -> List[Any]: """simple docstring""" _UpperCamelCase = False _UpperCamelCase = [] def __UpperCAmelCase ( self : str) -> Optional[Any]: """simple docstring""" if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq) def __UpperCAmelCase ( self : str , lowercase_ : int=False) -> List[Any]: """simple docstring""" _UpperCamelCase = DisjunctiveConstraint(self.token_ids) if stateful: _UpperCamelCase = self.seqlen _UpperCamelCase = self.current_seq _UpperCamelCase = self.completed return new_constraint class _UpperCAmelCase : '''simple docstring''' def __init__( self : Tuple , lowercase_ : List[Constraint]) -> Dict: """simple docstring""" _UpperCamelCase = constraints # max # of steps required to fulfill a given constraint _UpperCamelCase = max([c.seqlen for c in constraints]) _UpperCamelCase = len(__UpperCamelCase) _UpperCamelCase = False self.init_state() def __UpperCAmelCase ( self : Tuple) -> Optional[int]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = None _UpperCamelCase = [constraint.copy(stateful=__UpperCamelCase) for constraint in self.constraints] def __UpperCAmelCase ( self : List[Any]) -> Any: """simple docstring""" _UpperCamelCase = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints) * self.max_seqlen) + add def __UpperCAmelCase ( self : List[str]) -> Dict: """simple docstring""" _UpperCamelCase = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" _UpperCamelCase = constraint.advance() if isinstance(__UpperCamelCase , __UpperCamelCase): token_list.append(__UpperCamelCase) elif isinstance(__UpperCamelCase , __UpperCamelCase): token_list.extend(__UpperCamelCase) else: _UpperCamelCase = self.inprogress_constraint.advance() if isinstance(__UpperCamelCase , __UpperCamelCase): token_list.append(__UpperCamelCase) elif isinstance(__UpperCamelCase , __UpperCamelCase): token_list.extend(__UpperCamelCase) if len(__UpperCamelCase) == 0: return None else: return token_list def __UpperCAmelCase ( self : int , lowercase_ : Optional[List[int]]) -> Union[str, Any]: """simple docstring""" self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint _UpperCamelCase , _UpperCamelCase = self.add(__UpperCamelCase) # the entire list of constraints are fulfilled if self.completed: break def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int) -> Dict: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase): raise ValueError(f'`token_id` should be an `int`, but is `{token_id}`.') _UpperCamelCase , _UpperCamelCase = False, False if self.completed: _UpperCamelCase = True _UpperCamelCase = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.inprogress_constraint.update(__UpperCamelCase) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=__UpperCamelCase)) _UpperCamelCase = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint) _UpperCamelCase = None if len(self.pending_constraints) == 0: # we're done! _UpperCamelCase = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints): if pending_constraint.does_advance(__UpperCamelCase): _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = pending_constraint.update(__UpperCamelCase) if not stepped: raise Exception( "`constraint.update(token_id)` is not yielding incremental progress, " "even though `constraint.does_advance(token_id)` is true.") if complete: self.complete_constraints.append(__UpperCamelCase) _UpperCamelCase = None if not complete and stepped: _UpperCamelCase = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". _UpperCamelCase = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. _UpperCamelCase = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[Any]=True) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ConstraintListState(self.constraints) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: _UpperCamelCase = [ constraint.copy(stateful=__UpperCamelCase) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: _UpperCamelCase = self.inprogress_constraint.copy(stateful=__UpperCamelCase) _UpperCamelCase = [constraint.copy() for constraint in self.pending_constraints] return new_state
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.8.0''', '''To fix: pip install -r examples/pytorch/image-classification/requirements.txt''') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' with open(_UpperCamelCase , "rb" ) as f: _UpperCamelCase = Image.open(_UpperCamelCase ) return im.convert("RGB" ) @dataclass class _UpperCAmelCase : '''simple docstring''' __A = field( default=lowerCAmelCase, metadata={ '''help''': '''Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub).''' }, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) __A = field(default=lowerCAmelCase, metadata={'''help''': '''A folder containing the training data.'''} ) __A = field(default=lowerCAmelCase, metadata={'''help''': '''A folder containing the validation data.'''} ) __A = field( default=0.15, metadata={'''help''': '''Percent to split off of train for validation.'''} ) __A = field( default=lowerCAmelCase, metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) }, ) __A = field( default=lowerCAmelCase, metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) }, ) def __UpperCAmelCase ( self : Optional[int]) -> Union[str, Any]: """simple docstring""" if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory.") @dataclass class _UpperCAmelCase : '''simple docstring''' __A = field( default='''google/vit-base-patch16-224-in21k''', metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''}, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(lowerCAmelCase )}, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Where do you want to store the pretrained models downloaded from s3'''} ) __A = field( default='''main''', metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''}, ) __A = field(default=lowerCAmelCase, metadata={'''help''': '''Name or path of preprocessor config.'''} ) __A = field( default=lowerCAmelCase, metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) }, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Will enable to load a pretrained model whose head dimensions are different.'''}, ) def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = torch.stack([example["pixel_values"] for example in examples] ) _UpperCamelCase = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , _UpperCamelCase , _UpperCamelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCamelCase = training_args.get_process_log_level() logger.setLevel(_UpperCamelCase ) transformers.utils.logging.set_verbosity(_UpperCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCamelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCamelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCamelCase = {} if data_args.train_dir is not None: _UpperCamelCase = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCamelCase = os.path.join(data_args.validation_dir , "**" ) _UpperCamelCase = load_dataset( "imagefolder" , data_files=_UpperCamelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCamelCase = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _UpperCamelCase ) and data_args.train_val_split > 0.0: _UpperCamelCase = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCamelCase = split["train"] _UpperCamelCase = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCamelCase = dataset["train"].features["labels"].names _UpperCamelCase , _UpperCamelCase = {}, {} for i, label in enumerate(_UpperCamelCase ): _UpperCamelCase = str(_UpperCamelCase ) _UpperCamelCase = label # Load the accuracy metric from the datasets package _UpperCamelCase = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(a__ ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCamelCase = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(_UpperCamelCase ) , labelaid=_UpperCamelCase , idalabel=_UpperCamelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCamelCase = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=_UpperCamelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCamelCase = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCamelCase = image_processor.size["shortest_edge"] else: _UpperCamelCase = (image_processor.size["height"], image_processor.size["width"]) _UpperCamelCase = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCamelCase = Compose( [ RandomResizedCrop(_UpperCamelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCamelCase = Compose( [ Resize(_UpperCamelCase ), CenterCrop(_UpperCamelCase ), ToTensor(), normalize, ] ) def train_transforms(a__ ): _UpperCamelCase = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(a__ ): _UpperCamelCase = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCamelCase = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(_UpperCamelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCamelCase = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(_UpperCamelCase ) # Initalize our trainer _UpperCamelCase = Trainer( model=_UpperCamelCase , args=_UpperCamelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=_UpperCamelCase , tokenizer=_UpperCamelCase , data_collator=_UpperCamelCase , ) # Training if training_args.do_train: _UpperCamelCase = None if training_args.resume_from_checkpoint is not None: _UpperCamelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCamelCase = last_checkpoint _UpperCamelCase = trainer.train(resume_from_checkpoint=_UpperCamelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCamelCase = trainer.evaluate() trainer.log_metrics("eval" , _UpperCamelCase ) trainer.save_metrics("eval" , _UpperCamelCase ) # Write model card and (optionally) push to hub _UpperCamelCase = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**_UpperCamelCase ) else: trainer.create_model_card(**_UpperCamelCase ) if __name__ == "__main__": main()
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json""" ), }, } lowerCamelCase__ = { """yjernite/retribert-base-uncased""": 512, } lowerCamelCase__ = { """yjernite/retribert-base-uncased""": {"""do_lower_case""": True}, } class _UpperCAmelCase ( __UpperCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = PRETRAINED_INIT_CONFIGURATION __A = RetriBertTokenizer __A = ['''input_ids''', '''attention_mask'''] def __init__( self : int , lowercase_ : List[str]=None , lowercase_ : Optional[int]=None , lowercase_ : List[str]=True , lowercase_ : Tuple="[UNK]" , lowercase_ : int="[SEP]" , lowercase_ : Optional[Any]="[PAD]" , lowercase_ : Tuple="[CLS]" , lowercase_ : Dict="[MASK]" , lowercase_ : Optional[int]=True , lowercase_ : Union[str, Any]=None , **lowercase_ : List[str] , ) -> Dict: """simple docstring""" super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , tokenize_chinese_chars=lowercase_ , strip_accents=lowercase_ , **lowercase_ , ) _UpperCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get("lowercase" , lowercase_) != do_lower_case or normalizer_state.get("strip_accents" , lowercase_) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowercase_) != tokenize_chinese_chars ): _UpperCamelCase = getattr(lowercase_ , normalizer_state.pop("type")) _UpperCamelCase = do_lower_case _UpperCamelCase = strip_accents _UpperCamelCase = tokenize_chinese_chars _UpperCamelCase = normalizer_class(**lowercase_) _UpperCamelCase = do_lower_case def __UpperCAmelCase ( self : Any , lowercase_ : Optional[Any] , lowercase_ : List[Any]=None) -> str: """simple docstring""" _UpperCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self : str , lowercase_ : List[str] , lowercase_ : int = None) -> Tuple: """simple docstring""" _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def __UpperCAmelCase ( self : str , lowercase_ : str , lowercase_ : Optional[int] = None) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self._tokenizer.model.save(lowercase_ , name=lowercase_) return tuple(lowercase_)
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=__lowerCAmelCase ): '''simple docstring''' __A = ["note_seq"] def __init__( self : str , *lowercase_ : int , **lowercase_ : int) -> Tuple: """simple docstring""" requires_backends(self , ["note_seq"]) @classmethod def __UpperCAmelCase ( cls : Tuple , *lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" requires_backends(cls , ["note_seq"]) @classmethod def __UpperCAmelCase ( cls : Tuple , *lowercase_ : Any , **lowercase_ : Any) -> List[Any]: """simple docstring""" requires_backends(cls , ["note_seq"])
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class _UpperCAmelCase : '''simple docstring''' def __init__( self : str , lowercase_ : List[str] , lowercase_ : int=13 , lowercase_ : Dict=10 , lowercase_ : List[Any]=3 , lowercase_ : Dict=2 , lowercase_ : Tuple=2 , lowercase_ : int=2 , lowercase_ : Tuple=True , lowercase_ : Optional[int]=True , lowercase_ : Tuple=32 , lowercase_ : Any=5 , lowercase_ : Tuple=4 , lowercase_ : Union[str, Any]=37 , lowercase_ : int="gelu" , lowercase_ : Union[str, Any]=0.1 , lowercase_ : str=0.1 , lowercase_ : int=10 , lowercase_ : Dict=0.02 , lowercase_ : Any=0.9 , lowercase_ : Dict=None , ) -> Optional[int]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = num_channels _UpperCamelCase = patch_size _UpperCamelCase = tubelet_size _UpperCamelCase = num_frames _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = mask_ratio _UpperCamelCase = scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame _UpperCamelCase = (image_size // patch_size) ** 2 _UpperCamelCase = (num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos _UpperCamelCase = int(mask_ratio * self.seq_length) def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : Dict) -> List[str]: """simple docstring""" return VideoMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , tubelet_size=self.tubelet_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[Any] , lowercase_ : Any) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = VideoMAEModel(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = VideoMAEForPreTraining(lowercase_) model.to(lowercase_) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch _UpperCamelCase = torch.ones((self.num_masks,)) _UpperCamelCase = torch.cat([mask, torch.zeros(self.seq_length - mask.size(0))]) _UpperCamelCase = mask.expand(self.batch_size , -1).bool() _UpperCamelCase = model(lowercase_ , lowercase_) # model only returns predictions for masked patches _UpperCamelCase = mask.sum().item() _UpperCamelCase = 3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_masked_patches, decoder_num_labels)) def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase = config_and_inputs _UpperCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( UpperCamelCase_, UpperCamelCase_, unittest.TestCase ): '''simple docstring''' __A = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) __A = ( {'''feature-extraction''': VideoMAEModel, '''video-classification''': VideoMAEForVideoClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __UpperCAmelCase ( self : Optional[int]) -> Optional[int]: """simple docstring""" _UpperCamelCase = VideoMAEModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Optional[int] , lowercase_ : List[Any] , lowercase_ : List[Any]=False) -> List[str]: """simple docstring""" _UpperCamelCase = copy.deepcopy(lowercase_) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch _UpperCamelCase = torch.ones((self.model_tester.num_masks,)) _UpperCamelCase = torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0))]) _UpperCamelCase = mask.expand(self.model_tester.batch_size , -1).bool() _UpperCamelCase = bool_masked_pos.to(lowercase_) if return_labels: if model_class in [ *get_values(lowercase_), ]: _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase_) return inputs_dict def __UpperCAmelCase ( self : Optional[int]) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="VideoMAE does not use inputs_embeds") def __UpperCAmelCase ( self : int) -> List[str]: """simple docstring""" pass def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , nn.Linear)) def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowercase_) def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_) def __UpperCAmelCase ( self : Dict) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase_) @slow def __UpperCAmelCase ( self : Union[str, Any]) -> Optional[Any]: """simple docstring""" for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = VideoMAEModel.from_pretrained(lowercase_) self.assertIsNotNone(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> int: """simple docstring""" if not self.has_attentions: pass else: _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True for model_class in self.all_model_classes: _UpperCamelCase = self.model_tester.seq_length - self.model_tester.num_masks _UpperCamelCase = ( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.attentions self.assertEqual(len(lowercase_) , self.model_tester.num_hidden_layers) # check that output_attentions also work using config del inputs_dict["output_attentions"] _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.attentions self.assertEqual(len(lowercase_) , self.model_tester.num_hidden_layers) self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) _UpperCamelCase = len(lowercase_) # Check attention is always last and order is fine _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) self.assertEqual(out_len + 1 , len(lowercase_)) _UpperCamelCase = outputs.attentions self.assertEqual(len(lowercase_) , self.model_tester.num_hidden_layers) self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def __UpperCAmelCase ( self : Dict) -> Tuple: """simple docstring""" def check_hidden_states_output(lowercase_ : Any , lowercase_ : str , lowercase_ : List[Any]): _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.hidden_states _UpperCamelCase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowercase_) , lowercase_) _UpperCamelCase = self.model_tester.seq_length - self.model_tester.num_masks _UpperCamelCase = num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:]) , [seq_length, self.model_tester.hidden_size] , ) _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests.") def __UpperCAmelCase ( self : List[str]) -> str: """simple docstring""" pass def lowerCAmelCase__ ( ) ->int: '''simple docstring''' _UpperCamelCase = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename="eating_spaghetti.npy" , repo_type="dataset" ) _UpperCamelCase = np.load(snake_case_ ) return list(snake_case_ ) @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __UpperCAmelCase ( self : Any) -> List[str]: """simple docstring""" return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5]) if is_vision_available() else None ) @slow def __UpperCAmelCase ( self : Optional[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = VideoMAEForVideoClassification.from_pretrained("MCG-NJU/videomae-base-finetuned-kinetics").to( lowercase_) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_video() _UpperCamelCase = image_processor(lowercase_ , return_tensors="pt").to(lowercase_) # forward pass with torch.no_grad(): _UpperCamelCase = model(**lowercase_) # verify the logits _UpperCamelCase = torch.Size((1, 400)) self.assertEqual(outputs.logits.shape , lowercase_) _UpperCamelCase = torch.tensor([0.36_69, -0.06_88, -0.24_21]).to(lowercase_) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1e-4)) @slow def __UpperCAmelCase ( self : List[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = VideoMAEForPreTraining.from_pretrained("MCG-NJU/videomae-base-short").to(lowercase_) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_video() _UpperCamelCase = image_processor(lowercase_ , return_tensors="pt").to(lowercase_) # add boolean mask, indicating which patches to mask _UpperCamelCase = hf_hub_download(repo_id="hf-internal-testing/bool-masked-pos" , filename="bool_masked_pos.pt") _UpperCamelCase = torch.load(lowercase_) # forward pass with torch.no_grad(): _UpperCamelCase = model(**lowercase_) # verify the logits _UpperCamelCase = torch.Size([1, 1408, 1536]) _UpperCamelCase = torch.tensor( [[0.79_94, 0.96_12, 0.85_08], [0.74_01, 0.89_58, 0.83_02], [0.58_62, 0.74_68, 0.73_25]] , device=lowercase_) self.assertEqual(outputs.logits.shape , lowercase_) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , lowercase_ , atol=1e-4)) # verify the loss (`config.norm_pix_loss` = `True`) _UpperCamelCase = torch.tensor([0.51_42] , device=lowercase_) self.assertTrue(torch.allclose(outputs.loss , lowercase_ , atol=1e-4)) # verify the loss (`config.norm_pix_loss` = `False`) _UpperCamelCase = VideoMAEForPreTraining.from_pretrained("MCG-NJU/videomae-base-short" , norm_pix_loss=lowercase_).to( lowercase_) with torch.no_grad(): _UpperCamelCase = model(**lowercase_) _UpperCamelCase = torch.tensor(torch.tensor([0.64_69]) , device=lowercase_) self.assertTrue(torch.allclose(outputs.loss , lowercase_ , atol=1e-4))
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
def lowerCAmelCase__ ( a__ ) ->Optional[Any]: '''simple docstring''' if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError("multiplicative_persistence() only accepts integral values" ) if num < 0: raise ValueError("multiplicative_persistence() does not accept negative values" ) _UpperCamelCase = 0 _UpperCamelCase = str(UpperCamelCase__ ) while len(UpperCamelCase__ ) != 1: _UpperCamelCase = [int(UpperCamelCase__ ) for i in num_string] _UpperCamelCase = 1 for i in range(0 , len(UpperCamelCase__ ) ): total *= numbers[i] _UpperCamelCase = str(UpperCamelCase__ ) steps += 1 return steps def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError("additive_persistence() only accepts integral values" ) if num < 0: raise ValueError("additive_persistence() does not accept negative values" ) _UpperCamelCase = 0 _UpperCamelCase = str(UpperCamelCase__ ) while len(UpperCamelCase__ ) != 1: _UpperCamelCase = [int(UpperCamelCase__ ) for i in num_string] _UpperCamelCase = 0 for i in range(0 , len(UpperCamelCase__ ) ): total += numbers[i] _UpperCamelCase = str(UpperCamelCase__ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification lowerCamelCase__ = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co lowerCamelCase__ = '''main''' # Default branch name lowerCamelCase__ = '''f2c752cfc5c0ab6f4bdec59acea69eefbee381c2''' # One particular commit (not the top of `main`) lowerCamelCase__ = '''aaaaaaa''' # This commit does not exist, so we should 404. lowerCamelCase__ = '''d9e9f15bc825e4b2c9249e9578f884bbcb5e3684''' # Sha-1 of config.json on the top of `main`, for checking purposes lowerCamelCase__ = '''4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3''' @contextlib.contextmanager def lowerCAmelCase__ ( ) ->str: '''simple docstring''' print("Welcome!" ) yield print("Bye!" ) @contextlib.contextmanager def lowerCAmelCase__ ( ) ->List[str]: '''simple docstring''' print("Bonjour!" ) yield print("Au revoir!" ) class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> str: """simple docstring""" assert transformers.__spec__ is not None assert importlib.util.find_spec("transformers") is not None class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO) def __UpperCAmelCase ( self : str , lowercase_ : str) -> Dict: """simple docstring""" with ContextManagers([]): print("Transformers are awesome!") # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , "Transformers are awesome!\n") @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO) def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with ContextManagers([context_en()]): print("Transformers are awesome!") # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , "Welcome!\nTransformers are awesome!\nBye!\n") @unittest.mock.patch("sys.stdout" , new_callable=io.StringIO) def __UpperCAmelCase ( self : List[str] , lowercase_ : Tuple) -> str: """simple docstring""" with ContextManagers([context_fr(), context_en()]): print("Transformers are awesome!") # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , "Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n") @require_torch def __UpperCAmelCase ( self : Optional[int]) -> int: """simple docstring""" self.assertEqual(find_labels(lowercase_) , ["labels"]) self.assertEqual(find_labels(lowercase_) , ["labels", "next_sentence_label"]) self.assertEqual(find_labels(lowercase_) , ["start_positions", "end_positions"]) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' pass self.assertEqual(find_labels(lowercase_) , ["labels"]) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> str: """simple docstring""" self.assertEqual(find_labels(lowercase_) , ["labels"]) self.assertEqual(find_labels(lowercase_) , ["labels", "next_sentence_label"]) self.assertEqual(find_labels(lowercase_) , ["start_positions", "end_positions"]) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' pass self.assertEqual(find_labels(lowercase_) , ["labels"]) @require_flax def __UpperCAmelCase ( self : int) -> List[str]: """simple docstring""" self.assertEqual(find_labels(lowercase_) , []) self.assertEqual(find_labels(lowercase_) , []) self.assertEqual(find_labels(lowercase_) , []) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' pass self.assertEqual(find_labels(lowercase_) , [])
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Any: '''simple docstring''' _UpperCamelCase = to_pil_image(a__ ) _UpperCamelCase , _UpperCamelCase = pil_image.size _UpperCamelCase = pytesseract.image_to_data(a__ , lang=a__ , output_type="dict" , config=a__ ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates _UpperCamelCase = [idx for idx, word in enumerate(a__ ) if not word.strip()] _UpperCamelCase = [word for idx, word in enumerate(a__ ) if idx not in irrelevant_indices] _UpperCamelCase = [coord for idx, coord in enumerate(a__ ) if idx not in irrelevant_indices] _UpperCamelCase = [coord for idx, coord in enumerate(a__ ) if idx not in irrelevant_indices] _UpperCamelCase = [coord for idx, coord in enumerate(a__ ) if idx not in irrelevant_indices] _UpperCamelCase = [coord for idx, coord in enumerate(a__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format _UpperCamelCase = [] for x, y, w, h in zip(a__ , a__ , a__ , a__ ): _UpperCamelCase = [x, y, x + w, y + h] actual_boxes.append(a__ ) # finally, normalize the bounding boxes _UpperCamelCase = [] for box in actual_boxes: normalized_boxes.append(normalize_box(a__ , a__ , a__ ) ) assert len(a__ ) == len(a__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class _UpperCAmelCase ( snake_case__ ): '''simple docstring''' __A = ['''pixel_values'''] def __init__( self : Dict , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : bool = True , lowercase_ : float = 1 / 255 , lowercase_ : bool = True , lowercase_ : Union[float, Iterable[float]] = None , lowercase_ : Union[float, Iterable[float]] = None , lowercase_ : bool = True , lowercase_ : Optional[str] = None , lowercase_ : Optional[str] = "" , **lowercase_ : List[str] , ) -> None: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) _UpperCamelCase = size if size is not None else {"height": 224, "width": 224} _UpperCamelCase = get_size_dict(_SCREAMING_SNAKE_CASE) _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = resample _UpperCamelCase = do_rescale _UpperCamelCase = rescale_value _UpperCamelCase = do_normalize _UpperCamelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _UpperCamelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD _UpperCamelCase = apply_ocr _UpperCamelCase = ocr_lang _UpperCamelCase = tesseract_config def __UpperCAmelCase ( self : Any , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Any , ) -> np.ndarray: """simple docstring""" _UpperCamelCase = get_size_dict(_SCREAMING_SNAKE_CASE) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}') _UpperCamelCase = (size["height"], size["width"]) return resize(_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def __UpperCAmelCase ( self : int , lowercase_ : np.ndarray , lowercase_ : Union[int, float] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Union[str, Any] , ) -> np.ndarray: """simple docstring""" return rescale(_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : np.ndarray , lowercase_ : Union[float, Iterable[float]] , lowercase_ : Union[float, Iterable[float]] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Any , ) -> np.ndarray: """simple docstring""" return normalize(_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : ImageInput , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : List[Any]=None , lowercase_ : bool = None , lowercase_ : float = None , lowercase_ : bool = None , lowercase_ : Union[float, Iterable[float]] = None , lowercase_ : Union[float, Iterable[float]] = None , lowercase_ : bool = None , lowercase_ : Optional[str] = None , lowercase_ : Optional[str] = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : ChannelDimension = ChannelDimension.FIRST , **lowercase_ : Union[str, Any] , ) -> PIL.Image.Image: """simple docstring""" _UpperCamelCase = do_resize if do_resize is not None else self.do_resize _UpperCamelCase = size if size is not None else self.size _UpperCamelCase = get_size_dict(_SCREAMING_SNAKE_CASE) _UpperCamelCase = resample if resample is not None else self.resample _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_normalize if do_normalize is not None else self.do_normalize _UpperCamelCase = image_mean if image_mean is not None else self.image_mean _UpperCamelCase = image_std if image_std is not None else self.image_std _UpperCamelCase = apply_ocr if apply_ocr is not None else self.apply_ocr _UpperCamelCase = ocr_lang if ocr_lang is not None else self.ocr_lang _UpperCamelCase = tesseract_config if tesseract_config is not None else self.tesseract_config _UpperCamelCase = make_list_of_images(_SCREAMING_SNAKE_CASE) if not valid_images(_SCREAMING_SNAKE_CASE): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_normalize and (image_mean is None or image_std is None): raise ValueError("If do_normalize is True, image_mean and image_std must be specified.") # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(_SCREAMING_SNAKE_CASE) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , "pytesseract") _UpperCamelCase = [] _UpperCamelCase = [] for image in images: _UpperCamelCase , _UpperCamelCase = apply_tesseract(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) words_batch.append(_SCREAMING_SNAKE_CASE) boxes_batch.append(_SCREAMING_SNAKE_CASE) if do_resize: _UpperCamelCase = [self.resize(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE) for image in images] if do_normalize: _UpperCamelCase = [self.normalize(image=_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE) for image in images] _UpperCamelCase = [to_channel_dimension_format(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) for image in images] _UpperCamelCase = BatchFeature(data={"pixel_values": images} , tensor_type=_SCREAMING_SNAKE_CASE) if apply_ocr: _UpperCamelCase = words_batch _UpperCamelCase = boxes_batch return data
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
700
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
82
0
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
701
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = MODEL_FOR_MASKED_LM_MAPPING __A = TF_MODEL_FOR_MASKED_LM_MAPPING def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-0_5, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Patrick", "score": 2.1e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-0_5, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCamelCase = unmasker("My name is <mask> <mask>" , top_k=2) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt") # convert model to fp16 pipe.model.half() _UpperCamelCase = pipe("Paris is the [MASK] of France.") # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_) @slow @require_torch def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt") self.run_large_test(lowercase_) @slow @require_tf def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf") self.run_large_test(lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> Any: """simple docstring""" _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is John", "score": 0.0_08, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_07, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_51, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_14, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is Patrick", "score": 0.0_05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_00, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_00, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int]) -> int: """simple docstring""" if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)") _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = fill_masker.tokenizer _UpperCamelCase = fill_masker.model _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , ) with self.assertRaises(lowercase_): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase_): fill_masker("This is") self.run_test_top_k(lowercase_ , lowercase_) self.run_test_targets(lowercase_ , lowercase_) self.run_test_top_k_targets(lowercase_ , lowercase_) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Dict , lowercase_ : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = sorted(vocab.keys())[:2] # Pipeline argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Call argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Score equivalence _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCamelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_) == set(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) # Raises with invalid with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""]) with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets="") def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : str , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) # top_k=2, ntargets=3 _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase_) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCamelCase = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_).issubset(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase_) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCamelCase = fill_masker(f'My name is {tokenizer.mask_token}' , targets=lowercase_ , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_) , 3) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , )
82
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
702
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
0
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCamelCase__ ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 255.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
703
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
82
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCamelCase__ = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
704
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
0
# This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = multiprocessing.Manager() _UpperCamelCase = manager.list() _UpperCamelCase = multiprocessing.Process(target=a__ , args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append("timed out" ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Tuple: '''simple docstring''' with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil _UpperCamelCase = shutil.rmtree _UpperCamelCase = os.rmdir _UpperCamelCase = os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: _UpperCamelCase = {} with swallow_io(): with time_limit(a__ ): exec(a__ , a__ ) result.append("passed" ) except TimeoutException: result.append("timed out" ) except BaseException as e: result.append(f'failed: {e}' ) # Needed for cleaning up. _UpperCamelCase = rmtree _UpperCamelCase = rmdir _UpperCamelCase = chdir @contextlib.contextmanager def lowerCAmelCase__ ( a__ ) ->List[str]: '''simple docstring''' def signal_handler(a__ , a__ ): raise TimeoutException("Timed out!" ) signal.setitimer(signal.ITIMER_REAL , a__ ) signal.signal(signal.SIGALRM , a__ ) try: yield finally: signal.setitimer(signal.ITIMER_REAL , 0 ) @contextlib.contextmanager def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = WriteOnlyStringIO() with contextlib.redirect_stdout(a__ ): with contextlib.redirect_stderr(a__ ): with redirect_stdin(a__ ): yield @contextlib.contextmanager def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' with tempfile.TemporaryDirectory() as dirname: with chdir(a__ ): yield dirname class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' pass class _UpperCAmelCase ( io.StringIO ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[int] , *lowercase_ : int , **lowercase_ : Optional[int]) -> Any: """simple docstring""" raise OSError def __UpperCAmelCase ( self : List[str] , *lowercase_ : List[Any] , **lowercase_ : str) -> Optional[Any]: """simple docstring""" raise OSError def __UpperCAmelCase ( self : Dict , *lowercase_ : List[Any] , **lowercase_ : Any) -> Optional[int]: """simple docstring""" raise OSError def __UpperCAmelCase ( self : Optional[int] , *lowercase_ : List[str] , **lowercase_ : List[str]) -> List[str]: """simple docstring""" return False class _UpperCAmelCase ( contextlib._RedirectStream ): # type: ignore '''simple docstring''' __A = '''stdin''' @contextlib.contextmanager def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' if root == ".": yield return _UpperCamelCase = os.getcwd() os.chdir(a__ ) try: yield except BaseException as exc: raise exc finally: os.chdir(a__ ) def lowerCAmelCase__ ( a__=None ) ->Dict: '''simple docstring''' if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS , (maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA , (maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK , (maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins _UpperCamelCase = None _UpperCamelCase = None import os _UpperCamelCase = "1" _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None import shutil _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None import subprocess _UpperCamelCase = None # type: ignore _UpperCamelCase = None import sys _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None
705
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
82
0
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets lowerCamelCase__ = '''\ @inproceedings{popovic-2015-chrf, title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation", month = sep, year = "2015", address = "Lisbon, Portugal", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W15-3049", doi = "10.18653/v1/W15-3049", pages = "392--395", } @inproceedings{popovic-2017-chrf, title = "chr{F}++: words helping character n-grams", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Second Conference on Machine Translation", month = sep, year = "2017", address = "Copenhagen, Denmark", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W17-4770", doi = "10.18653/v1/W17-4770", pages = "612--618", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' lowerCamelCase__ = '''\ ChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches, and ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation that is already present in sacrebleu. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information. ''' lowerCamelCase__ = ''' Produces ChrF(++) scores for hypotheses given reference translations. Args: predictions (list of str): The predicted sentences. references (list of list of str): The references. There should be one reference sub-list for each prediction sentence. char_order (int): Character n-gram order. Defaults to `6`. word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`. beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`. lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`. whitespace (bool): If `True`, include whitespaces when extracting character n-grams. eps_smoothing (bool): If `True`, applies epsilon smoothing similar to reference chrF++.py, NLTK and Moses implementations. If `False`, it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`. Returns: \'score\' (float): The chrF (chrF++) score, \'char_order\' (int): The character n-gram order, \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++, \'beta\' (int): Determine the importance of recall w.r.t precision Examples: Example 1--a simple example of calculating chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, references=reference) >>> print(results) {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2} Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2) >>> print(results) {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2, ... lowercase=True) >>> print(results) {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : str) -> str: """simple docstring""" if version.parse(scb.__version__) < version.parse("1.4.12"): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`.") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/mjpost/sacreBLEU#chrf--chrf" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence"), "references": datasets.Sequence(datasets.Value("string" , id="sequence") , id="references"), }) , codebase_urls=["https://github.com/mjpost/sacreBLEU#chrf--chrf"] , reference_urls=[ "https://github.com/m-popovic/chrF", ] , ) def __UpperCAmelCase ( self : Any , lowercase_ : List[str] , lowercase_ : int , lowercase_ : int = CHRF.CHAR_ORDER , lowercase_ : int = CHRF.WORD_ORDER , lowercase_ : int = CHRF.BETA , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : bool = False , ) -> int: """simple docstring""" _UpperCamelCase = len(references[0]) if any(len(lowercase_) != references_per_prediction for refs in references): raise ValueError("Sacrebleu requires the same number of references for each prediction") _UpperCamelCase = [[refs[i] for refs in references] for i in range(lowercase_)] _UpperCamelCase = CHRF(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_) _UpperCamelCase = sb_chrf.corpus_score(lowercase_ , lowercase_) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
706
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
82
0
def lowerCAmelCase__ ( a__ ) ->bool: '''simple docstring''' _UpperCamelCase = [int(a__ ) for i in ip_va_address.split("." ) if i.isdigit()] return len(a__ ) == 4 and all(0 <= int(a__ ) <= 254 for octet in octets ) if __name__ == "__main__": lowerCamelCase__ = input().strip() lowerCamelCase__ = '''valid''' if is_ip_va_address_valid(ip) else '''invalid''' print(F"{ip} is a {valid_or_invalid} IP v4 address.")
707
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
82
0
import os from distutils.util import strtobool def lowerCAmelCase__ ( a__ , a__ ) ->str: '''simple docstring''' for e in env_keys: _UpperCamelCase = int(os.environ.get(a__ , -1 ) ) if val >= 0: return val return default def lowerCAmelCase__ ( a__ , a__=False ) ->str: '''simple docstring''' _UpperCamelCase = os.environ.get(a__ , str(a__ ) ) return strtobool(a__ ) == 1 # As its name indicates `strtobool` actually returns an int... def lowerCAmelCase__ ( a__ , a__="no" ) ->int: '''simple docstring''' _UpperCamelCase = os.environ.get(a__ , str(a__ ) ) return value
708
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
82
0
from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance lowerCamelCase__ = 6378137.0 lowerCamelCase__ = 6356752.314245 lowerCamelCase__ = 637_8137 def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->float: '''simple docstring''' _UpperCamelCase = (AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude _UpperCamelCase = atan((1 - flattening) * tan(radians(a__ ) ) ) _UpperCamelCase = atan((1 - flattening) * tan(radians(a__ ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius _UpperCamelCase = haversine_distance(a__ , a__ , a__ , a__ ) / EQUATORIAL_RADIUS # Intermediate P and Q values _UpperCamelCase = (b_lata + b_lata) / 2 _UpperCamelCase = (b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) _UpperCamelCase = (sin(a__ ) ** 2) * (cos(a__ ) ** 2) _UpperCamelCase = cos(sigma / 2 ) ** 2 _UpperCamelCase = (sigma - sin(a__ )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) _UpperCamelCase = (cos(a__ ) ** 2) * (sin(a__ ) ** 2) _UpperCamelCase = sin(sigma / 2 ) ** 2 _UpperCamelCase = (sigma + sin(a__ )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
709
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , *lowercase_ : List[str] , **lowercase_ : Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
82
0
import gc import threading import time import psutil import torch class _UpperCAmelCase : '''simple docstring''' def __init__( self : Union[str, Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = psutil.Process() _UpperCamelCase = False def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = -1 while True: _UpperCamelCase = max(self.process.memory_info().rss , self.cpu_memory_peak) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" _UpperCamelCase = True _UpperCamelCase = threading.Thread(target=self.peak_monitor) _UpperCamelCase = True self.thread.start() def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = False self.thread.join() return self.cpu_memory_peak lowerCamelCase__ = PeakCPUMemory() def lowerCAmelCase__ ( ) ->str: '''simple docstring''' _UpperCamelCase = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem _UpperCamelCase = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): _UpperCamelCase = torch.cuda.memory_allocated(a__ ) torch.cuda.reset_peak_memory_stats() return measures def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem _UpperCamelCase = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 _UpperCamelCase = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): _UpperCamelCase = (torch.cuda.memory_allocated(a__ ) - start_measures[str(a__ )]) / 2**20 _UpperCamelCase = (torch.cuda.max_memory_allocated(a__ ) - start_measures[str(a__ )]) / 2**20 return measures def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' print(f'{description}:' ) print(f'- Time: {measures["time"]:.2f}s' ) for i in range(torch.cuda.device_count() ): print(f'- GPU {i} allocated: {measures[str(a__ )]:.2f}MiB' ) _UpperCamelCase = measures[f'{i}-peak'] print(f'- GPU {i} peak: {peak:.2f}MiB' ) print(f'- CPU RAM allocated: {measures["cpu"]:.2f}MiB' ) print(f'- CPU RAM peak: {measures["cpu-peak"]:.2f}MiB' )
710
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
82
0
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCamelCase__ = (720, 1280) # Height, Width lowerCamelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCamelCase__ = 1 / 100 lowerCamelCase__ = '''''' lowerCamelCase__ = '''''' lowerCamelCase__ = '''''' lowerCamelCase__ = 250 def lowerCAmelCase__ ( ) ->None: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_dataset(a__ , a__ ) for index in range(a__ ): _UpperCamelCase = random.sample(range(len(a__ ) ) , 4 ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = update_image_and_anno( a__ , a__ , a__ , a__ , a__ , filter_scale=a__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _UpperCamelCase = random_chars(32 ) _UpperCamelCase = path.split(os.sep )[-1].rsplit("." , 1 )[0] _UpperCamelCase = f'{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}' cva.imwrite(f'{file_root}.jpg' , a__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}' ) _UpperCamelCase = [] for anno in new_annos: _UpperCamelCase = anno[3] - anno[1] _UpperCamelCase = anno[4] - anno[2] _UpperCamelCase = anno[1] + width / 2 _UpperCamelCase = anno[2] + height / 2 _UpperCamelCase = f'{anno[0]} {x_center} {y_center} {width} {height}' annos_list.append(a__ ) with open(f'{file_root}.txt' , "w" ) as outfile: outfile.write("\n".join(line for line in annos_list ) ) def lowerCAmelCase__ ( a__ , a__ ) ->tuple[list, list]: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] for label_file in glob.glob(os.path.join(a__ , "*.txt" ) ): _UpperCamelCase = label_file.split(os.sep )[-1].rsplit("." , 1 )[0] with open(a__ ) as in_file: _UpperCamelCase = in_file.readlines() _UpperCamelCase = os.path.join(a__ , f'{label_name}.jpg' ) _UpperCamelCase = [] for obj_list in obj_lists: _UpperCamelCase = obj_list.rstrip("\n" ).split(" " ) _UpperCamelCase = float(obj[1] ) - float(obj[3] ) / 2 _UpperCamelCase = float(obj[2] ) - float(obj[4] ) / 2 _UpperCamelCase = float(obj[1] ) + float(obj[3] ) / 2 _UpperCamelCase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(a__ ) labels.append(a__ ) return img_paths, labels def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ , a__ = 0.0 , ) ->tuple[list, list, str]: '''simple docstring''' _UpperCamelCase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) _UpperCamelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase = int(scale_x * output_size[1] ) _UpperCamelCase = int(scale_y * output_size[0] ) _UpperCamelCase = [] _UpperCamelCase = [] for i, index in enumerate(a__ ): _UpperCamelCase = all_img_list[index] path_list.append(a__ ) _UpperCamelCase = all_annos[index] _UpperCamelCase = cva.imread(a__ ) if i == 0: # top-left _UpperCamelCase = cva.resize(a__ , (divid_point_x, divid_point_y) ) _UpperCamelCase = img for bbox in img_annos: _UpperCamelCase = bbox[1] * scale_x _UpperCamelCase = bbox[2] * scale_y _UpperCamelCase = bbox[3] * scale_x _UpperCamelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _UpperCamelCase = cva.resize(a__ , (output_size[1] - divid_point_x, divid_point_y) ) _UpperCamelCase = img for bbox in img_annos: _UpperCamelCase = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase = bbox[2] * scale_y _UpperCamelCase = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _UpperCamelCase = cva.resize(a__ , (divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase = img for bbox in img_annos: _UpperCamelCase = bbox[1] * scale_x _UpperCamelCase = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase = bbox[3] * scale_x _UpperCamelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _UpperCamelCase = cva.resize( a__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase = img for bbox in img_annos: _UpperCamelCase = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _UpperCamelCase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" _UpperCamelCase = ascii_lowercase + digits return "".join(random.choice(a__ ) for _ in range(a__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
711
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
82
0
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[int] , lowercase_ : Optional[int] , lowercase_ : List[Any]=7 , lowercase_ : Optional[Any]=3 , lowercase_ : Any=18 , lowercase_ : Tuple=30 , lowercase_ : List[str]=400 , lowercase_ : List[Any]=True , lowercase_ : Optional[Any]=None , lowercase_ : List[Any]=True , lowercase_ : Optional[Any]=[0.5, 0.5, 0.5] , lowercase_ : List[str]=[0.5, 0.5, 0.5] , ) -> List[Any]: """simple docstring""" _UpperCamelCase = size if size is not None else {"height": 18, "width": 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_normalize _UpperCamelCase = image_mean _UpperCamelCase = image_std def __UpperCAmelCase ( self : int) -> Union[str, Any]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = DPTImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = DPTImageProcessingTester(self) @property def __UpperCAmelCase ( self : Union[str, Any]) -> Optional[int]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowercase_ , "image_mean")) self.assertTrue(hasattr(lowercase_ , "image_std")) self.assertTrue(hasattr(lowercase_ , "do_normalize")) self.assertTrue(hasattr(lowercase_ , "do_resize")) self.assertTrue(hasattr(lowercase_ , "size")) def __UpperCAmelCase ( self : Optional[int]) -> int: """simple docstring""" _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"height": 18, "width": 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"height": 42, "width": 42}) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched _UpperCamelCase = image_processing(lowercase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def __UpperCAmelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , numpify=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched _UpperCamelCase = image_processing(lowercase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def __UpperCAmelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , torchify=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched _UpperCamelCase = image_processing(lowercase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
712
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
82
0
import os import string import sys lowerCamelCase__ = 1 << 8 lowerCamelCase__ = { '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 27, '''up''': 65 + ARROW_KEY_FLAG, '''down''': 66 + ARROW_KEY_FLAG, '''right''': 67 + ARROW_KEY_FLAG, '''left''': 68 + ARROW_KEY_FLAG, '''mod_int''': 91, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 50, '''delete''': 51, '''pg_up''': 53, '''pg_down''': 54, } lowerCamelCase__ = KEYMAP['''up'''] lowerCamelCase__ = KEYMAP['''left'''] if sys.platform == "win32": lowerCamelCase__ = [] lowerCamelCase__ = { B'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, B'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, B'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, B'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, B'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, B'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, B'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, B'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(10): lowerCamelCase__ = ord(str(i)) def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' if os.name == "nt": import msvcrt _UpperCamelCase = "mbcs" # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(a__ ) == 0: # Read the keystroke _UpperCamelCase = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): _UpperCamelCase = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: _UpperCamelCase = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP["mod_int"] ) ) WIN_CH_BUFFER.append(a__ ) if ord(a__ ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) _UpperCamelCase = chr(KEYMAP["esc"] ) except KeyError: _UpperCamelCase = cha[1] else: _UpperCamelCase = ch.decode(a__ ) else: _UpperCamelCase = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty _UpperCamelCase = sys.stdin.fileno() _UpperCamelCase = termios.tcgetattr(a__ ) try: tty.setraw(a__ ) _UpperCamelCase = sys.stdin.read(1 ) finally: termios.tcsetattr(a__ , termios.TCSADRAIN , a__ ) return ch def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = get_raw_chars() if ord(a__ ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(a__ ) == KEYMAP["esc"]: _UpperCamelCase = get_raw_chars() if ord(a__ ) == KEYMAP["mod_int"]: _UpperCamelCase = get_raw_chars() if ord(a__ ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(a__ ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(a__ ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' if "resnet-50" in model_name: _UpperCamelCase = ResNetConfig.from_pretrained("microsoft/resnet-50" ) elif "resnet-101" in model_name: _UpperCamelCase = ResNetConfig.from_pretrained("microsoft/resnet-101" ) else: raise ValueError("Model name should include either resnet50 or resnet101" ) _UpperCamelCase = DetrConfig(use_timm_backbone=a__ , backbone_config=a__ ) # set label attributes _UpperCamelCase = "panoptic" in model_name if is_panoptic: _UpperCamelCase = 250 else: _UpperCamelCase = 91 _UpperCamelCase = "huggingface/label-files" _UpperCamelCase = "coco-detection-id2label.json" _UpperCamelCase = json.load(open(hf_hub_download(a__ , a__ , repo_type="dataset" ) , "r" ) ) _UpperCamelCase = {int(a__ ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight") ) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight") ) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias") ) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean") ) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var") ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean', ) ) rename_keys.append( ( f'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var', f'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight', ) ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias') ) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append( (f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight', ) ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) return rename_keys def lowerCAmelCase__ ( a__ , a__ , a__ ) ->List[Any]: '''simple docstring''' _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val def lowerCAmelCase__ ( a__ , a__=False ) ->List[Any]: '''simple docstring''' _UpperCamelCase = "" if is_panoptic: _UpperCamelCase = "detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _UpperCamelCase = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[:256, :] _UpperCamelCase = in_proj_bias[:256] _UpperCamelCase = in_proj_weight[256:512, :] _UpperCamelCase = in_proj_bias[256:512] _UpperCamelCase = in_proj_weight[-256:, :] _UpperCamelCase = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _UpperCamelCase = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _UpperCamelCase = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[:256, :] _UpperCamelCase = in_proj_bias[:256] _UpperCamelCase = in_proj_weight[256:512, :] _UpperCamelCase = in_proj_bias[256:512] _UpperCamelCase = in_proj_weight[-256:, :] _UpperCamelCase = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _UpperCamelCase = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _UpperCamelCase = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _UpperCamelCase = in_proj_weight_cross_attn[:256, :] _UpperCamelCase = in_proj_bias_cross_attn[:256] _UpperCamelCase = in_proj_weight_cross_attn[256:512, :] _UpperCamelCase = in_proj_bias_cross_attn[256:512] _UpperCamelCase = in_proj_weight_cross_attn[-256:, :] _UpperCamelCase = in_proj_bias_cross_attn[-256:] def lowerCAmelCase__ ( ) ->Tuple: '''simple docstring''' _UpperCamelCase = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCamelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( a__ , a__=None , a__=False ) ->Dict: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_detr_config(a__ ) # load original model from torch hub _UpperCamelCase = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f'Converting model {model_name}...' ) _UpperCamelCase = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=a__ ).eval() _UpperCamelCase = detr.state_dict() # rename keys for src, dest in create_rename_keys(a__ ): if is_panoptic: _UpperCamelCase = "detr." + src rename_key(a__ , a__ , a__ ) # query, key and value matrices need special treatment read_in_q_k_v(a__ , is_panoptic=a__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _UpperCamelCase = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val # finally, create HuggingFace model and load state dict _UpperCamelCase = DetrForSegmentation(a__ ) if is_panoptic else DetrForObjectDetection(a__ ) model.load_state_dict(a__ ) model.eval() # verify our conversion on an image _UpperCamelCase = "coco_panoptic" if is_panoptic else "coco_detection" _UpperCamelCase = DetrImageProcessor(format=a__ ) _UpperCamelCase = processor(images=prepare_img() , return_tensors="pt" ) _UpperCamelCase = encoding["pixel_values"] _UpperCamelCase = detr(a__ ) _UpperCamelCase = model(a__ ) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) processor.save_pretrained(a__ ) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub..." ) model.push_to_hub(f'nielsr/{model_name}' ) processor.push_to_hub(f'nielsr/{model_name}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''detr-resnet-50''', type=str, choices=['''detr-resnet-50''', '''detr-resnet-101'''], help='''Name of the DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Whether to push the model to the hub or not.''') lowerCamelCase__ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' if isinstance(a__ , a__ ): raise TypeError("'float' object cannot be interpreted as an integer" ) if isinstance(a__ , a__ ): raise TypeError("'str' object cannot be interpreted as an integer" ) if num == 0: return "0b0" _UpperCamelCase = False if num < 0: _UpperCamelCase = True _UpperCamelCase = -num _UpperCamelCase = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(a__ ) for e in binary ) return "0b" + "".join(str(a__ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( a__ , a__ = 0.0 , a__ = 1.0 ) ->int: '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html lowerCamelCase__ = '''platform''' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def lowerCAmelCase__ ( a__ , a__ , a__=None , a__=None , a__=None , a__=None , a__=None , a__=None , ) ->Union[str, Any]: '''simple docstring''' if attention_mask is None: _UpperCamelCase = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _UpperCamelCase = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _UpperCamelCase = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[str] , lowercase_ : int , lowercase_ : List[str]=13 , lowercase_ : Optional[int]=7 , lowercase_ : List[str]=True , lowercase_ : int=False , lowercase_ : List[str]=99 , lowercase_ : Any=16 , lowercase_ : Any=2 , lowercase_ : Union[str, Any]=4 , lowercase_ : List[Any]=4 , lowercase_ : Any="gelu" , lowercase_ : str=0.1 , lowercase_ : Optional[int]=0.1 , lowercase_ : List[str]=32 , lowercase_ : Optional[int]=2 , lowercase_ : List[str]=1 , lowercase_ : str=0 , lowercase_ : Optional[int]=0.02 , ) -> int: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id _UpperCamelCase = initializer_range def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" _UpperCamelCase = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) , 3 , self.vocab_size) _UpperCamelCase = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa)) , -1) _UpperCamelCase = shift_tokens_right(lowercase_ , 1 , 2) _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowercase_ , ) _UpperCamelCase = prepare_blenderbot_inputs_dict(lowercase_ , lowercase_ , lowercase_) return config, inputs_dict def __UpperCAmelCase ( self : Tuple) -> List[str]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.prepare_config_and_inputs() return config, inputs_dict def __UpperCAmelCase ( self : Tuple , lowercase_ : Dict , lowercase_ : List[str] , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 20 _UpperCamelCase = model_class_name(lowercase_) _UpperCamelCase = model.encode(inputs_dict["input_ids"]) _UpperCamelCase , _UpperCamelCase = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _UpperCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowercase_ , lowercase_) _UpperCamelCase = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4") _UpperCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _UpperCamelCase = model.decode( decoder_input_ids[:, :-1] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=lowercase_ , decoder_position_ids=lowercase_ , ) _UpperCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4") _UpperCamelCase = model.decode( decoder_input_ids[:, -1:] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowercase_ , ) _UpperCamelCase = model.decode(lowercase_ , lowercase_) _UpperCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Any , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = 20 _UpperCamelCase = model_class_name(lowercase_) _UpperCamelCase = model.encode(inputs_dict["input_ids"]) _UpperCamelCase , _UpperCamelCase = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _UpperCamelCase = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1])), ] , axis=-1 , ) _UpperCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowercase_ , lowercase_) _UpperCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _UpperCamelCase = model.decode( decoder_input_ids[:, :-1] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=lowercase_ , decoder_position_ids=lowercase_ , ) _UpperCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4") _UpperCamelCase = model.decode( decoder_input_ids[:, -1:] , lowercase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowercase_ , decoder_position_ids=lowercase_ , ) _UpperCamelCase = model.decode(lowercase_ , lowercase_ , decoder_attention_mask=lowercase_) _UpperCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') @require_flax class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = 99 def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _UpperCamelCase = input_ids.shape[0] _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def __UpperCAmelCase ( self : int) -> Optional[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self._get_config_and_data() _UpperCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowercase_) _UpperCamelCase = lm_model(input_ids=lowercase_) _UpperCamelCase = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["logits"].shape , lowercase_) def __UpperCAmelCase ( self : Tuple) -> List[str]: """simple docstring""" _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _UpperCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowercase_) _UpperCamelCase = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa) _UpperCamelCase = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa) _UpperCamelCase = lm_model(input_ids=lowercase_ , decoder_input_ids=lowercase_) _UpperCamelCase = (*summary.shape, config.vocab_size) self.assertEqual(outputs["logits"].shape , lowercase_) def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" _UpperCamelCase = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa) _UpperCamelCase = shift_tokens_right(lowercase_ , 1 , 2) _UpperCamelCase = np.equal(lowercase_ , 1).astype(np.floataa).sum() _UpperCamelCase = np.equal(lowercase_ , 1).astype(np.floataa).sum() self.assertEqual(shifted.shape , input_ids.shape) self.assertEqual(lowercase_ , n_pad_before - 1) self.assertTrue(np.equal(shifted[:, 0] , 2).all()) @require_flax class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase, lowerCAmelCase ): '''simple docstring''' __A = True __A = ( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) __A = (FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = FlaxBlenderbotSmallModelTester(self) def __UpperCAmelCase ( self : Dict) -> int: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowercase_ , lowercase_ , lowercase_) def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowercase_ , lowercase_ , lowercase_) def __UpperCAmelCase ( self : List[str]) -> str: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): _UpperCamelCase = self._prepare_for_class(lowercase_ , lowercase_) _UpperCamelCase = model_class(lowercase_) @jax.jit def encode_jitted(lowercase_ : Union[str, Any] , lowercase_ : Tuple=None , **lowercase_ : Any): return model.encode(input_ids=lowercase_ , attention_mask=lowercase_) with self.subTest("JIT Enabled"): _UpperCamelCase = encode_jitted(**lowercase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): _UpperCamelCase = encode_jitted(**lowercase_).to_tuple() self.assertEqual(len(lowercase_) , len(lowercase_)) for jitted_output, output in zip(lowercase_ , lowercase_): self.assertEqual(jitted_output.shape , output.shape) def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): _UpperCamelCase = model_class(lowercase_) _UpperCamelCase = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"]) _UpperCamelCase = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowercase_ : Any , lowercase_ : int , lowercase_ : List[str]): return model.decode( decoder_input_ids=lowercase_ , decoder_attention_mask=lowercase_ , encoder_outputs=lowercase_ , ) with self.subTest("JIT Enabled"): _UpperCamelCase = decode_jitted(**lowercase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): _UpperCamelCase = decode_jitted(**lowercase_).to_tuple() self.assertEqual(len(lowercase_) , len(lowercase_)) for jitted_output, output in zip(lowercase_ , lowercase_): self.assertEqual(jitted_output.shape , output.shape) @slow def __UpperCAmelCase ( self : Optional[int]) -> List[str]: """simple docstring""" for model_class_name in self.all_model_classes: _UpperCamelCase = model_class_name.from_pretrained("facebook/blenderbot_small-90M") # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _UpperCamelCase = np.ones((1, 1)) * model.config.eos_token_id _UpperCamelCase = model(lowercase_) self.assertIsNotNone(lowercase_)
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = field(default='''image-classification''', metadata={'''include_in_asdict_even_if_is_default''': True} ) __A = Features({'''image''': Image()} ) __A = Features({'''labels''': ClassLabel} ) __A = '''image''' __A = '''labels''' def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> List[str]: """simple docstring""" if self.label_column not in features: raise ValueError(f'Column {self.label_column} is not present in features.') if not isinstance(features[self.label_column] , lowercase_): raise ValueError(f'Column {self.label_column} is not a ClassLabel.') _UpperCamelCase = copy.deepcopy(self) _UpperCamelCase = self.label_schema.copy() _UpperCamelCase = features[self.label_column] _UpperCamelCase = label_schema return task_template @property def __UpperCAmelCase ( self : Dict) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position lowerCamelCase__ = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowerCamelCase__ = concatenate_datasets lowerCamelCase__ = DownloadConfig lowerCamelCase__ = DownloadManager lowerCamelCase__ = DownloadMode lowerCamelCase__ = DownloadConfig lowerCamelCase__ = DownloadMode lowerCamelCase__ = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = (CMStochasticIterativeScheduler,) __A = 10 def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> int: """simple docstring""" _UpperCamelCase = { "num_train_timesteps": 201, "sigma_min": 0.0_02, "sigma_max": 80.0, } config.update(**lowercase_) return config def __UpperCAmelCase ( self : Union[str, Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = 10 _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = self.scheduler_classes[0](**lowercase_) scheduler.set_timesteps(lowercase_) _UpperCamelCase = scheduler.timesteps[0] _UpperCamelCase = scheduler.timesteps[1] _UpperCamelCase = self.dummy_sample _UpperCamelCase = 0.1 * sample _UpperCamelCase = scheduler.step(lowercase_ , lowercase_ , lowercase_).prev_sample _UpperCamelCase = scheduler.step(lowercase_ , lowercase_ , lowercase_).prev_sample self.assertEqual(output_a.shape , sample.shape) self.assertEqual(output_a.shape , output_a.shape) def __UpperCAmelCase ( self : int) -> Optional[Any]: """simple docstring""" for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_) def __UpperCAmelCase ( self : Tuple) -> Optional[int]: """simple docstring""" for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=lowercase_) def __UpperCAmelCase ( self : List[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.scheduler_classes[0] _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = scheduler_class(**lowercase_) _UpperCamelCase = 1 scheduler.set_timesteps(lowercase_) _UpperCamelCase = scheduler.timesteps _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = self.dummy_model() _UpperCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(lowercase_): # 1. scale model input _UpperCamelCase = scheduler.scale_model_input(lowercase_ , lowercase_) # 2. predict noise residual _UpperCamelCase = model(lowercase_ , lowercase_) # 3. predict previous sample x_t-1 _UpperCamelCase = scheduler.step(lowercase_ , lowercase_ , lowercase_ , generator=lowercase_).prev_sample _UpperCamelCase = pred_prev_sample _UpperCamelCase = torch.sum(torch.abs(lowercase_)) _UpperCamelCase = torch.mean(torch.abs(lowercase_)) assert abs(result_sum.item() - 192.7614) < 1e-2 assert abs(result_mean.item() - 0.25_10) < 1e-3 def __UpperCAmelCase ( self : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = self.scheduler_classes[0] _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = scheduler_class(**lowercase_) _UpperCamelCase = [106, 0] scheduler.set_timesteps(timesteps=lowercase_) _UpperCamelCase = scheduler.timesteps _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = self.dummy_model() _UpperCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input _UpperCamelCase = scheduler.scale_model_input(lowercase_ , lowercase_) # 2. predict noise residual _UpperCamelCase = model(lowercase_ , lowercase_) # 3. predict previous sample x_t-1 _UpperCamelCase = scheduler.step(lowercase_ , lowercase_ , lowercase_ , generator=lowercase_).prev_sample _UpperCamelCase = pred_prev_sample _UpperCamelCase = torch.sum(torch.abs(lowercase_)) _UpperCamelCase = torch.mean(torch.abs(lowercase_)) assert abs(result_sum.item() - 347.6357) < 1e-2 assert abs(result_mean.item() - 0.45_27) < 1e-3 def __UpperCAmelCase ( self : Optional[Any]) -> Dict: """simple docstring""" _UpperCamelCase = self.scheduler_classes[0] _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = scheduler_class(**lowercase_) _UpperCamelCase = [39, 30, 12, 15, 0] with self.assertRaises(lowercase_ , msg="`timesteps` must be in descending order."): scheduler.set_timesteps(timesteps=lowercase_) def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" _UpperCamelCase = self.scheduler_classes[0] _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = scheduler_class(**lowercase_) _UpperCamelCase = [39, 30, 12, 1, 0] _UpperCamelCase = len(lowercase_) with self.assertRaises(lowercase_ , msg="Can only pass one of `num_inference_steps` or `timesteps`."): scheduler.set_timesteps(num_inference_steps=lowercase_ , timesteps=lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> Dict: """simple docstring""" _UpperCamelCase = self.scheduler_classes[0] _UpperCamelCase = self.get_scheduler_config() _UpperCamelCase = scheduler_class(**lowercase_) _UpperCamelCase = [scheduler.config.num_train_timesteps] with self.assertRaises( lowercase_ , msg="`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}" , ): scheduler.set_timesteps(timesteps=lowercase_)
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
import math def lowerCAmelCase__ ( a__ , a__ ) ->float: '''simple docstring''' return math.pow(a__ , 2 ) - a def lowerCAmelCase__ ( a__ ) ->float: '''simple docstring''' return 2 * x def lowerCAmelCase__ ( a__ ) ->float: '''simple docstring''' _UpperCamelCase = 2.0 while start <= a: _UpperCamelCase = math.pow(a__ , 2 ) return start def lowerCAmelCase__ ( a__ , a__ = 9_999 , a__ = 0.00000000000001 ) ->float: '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _UpperCamelCase = get_initial_point(a__ ) for _ in range(a__ ): _UpperCamelCase = value _UpperCamelCase = value - fx(a__ , a__ ) / fx_derivative(a__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
700
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
82
0
def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
701
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = MODEL_FOR_MASKED_LM_MAPPING __A = TF_MODEL_FOR_MASKED_LM_MAPPING def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-0_5, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Patrick", "score": 2.1e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-0_5, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCamelCase = unmasker("My name is <mask> <mask>" , top_k=2) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt") # convert model to fp16 pipe.model.half() _UpperCamelCase = pipe("Paris is the [MASK] of France.") # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_) @slow @require_torch def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt") self.run_large_test(lowercase_) @slow @require_tf def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf") self.run_large_test(lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> Any: """simple docstring""" _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is John", "score": 0.0_08, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_07, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_51, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_14, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is Patrick", "score": 0.0_05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_00, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_00, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int]) -> int: """simple docstring""" if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)") _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = fill_masker.tokenizer _UpperCamelCase = fill_masker.model _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , ) with self.assertRaises(lowercase_): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase_): fill_masker("This is") self.run_test_top_k(lowercase_ , lowercase_) self.run_test_targets(lowercase_ , lowercase_) self.run_test_top_k_targets(lowercase_ , lowercase_) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Dict , lowercase_ : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = sorted(vocab.keys())[:2] # Pipeline argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Call argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Score equivalence _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCamelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_) == set(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) # Raises with invalid with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""]) with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets="") def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : str , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) # top_k=2, ntargets=3 _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase_) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCamelCase = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_).issubset(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase_) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCamelCase = fill_masker(f'My name is {tokenizer.mask_token}' , targets=lowercase_ , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_) , 3) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , )
82
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''kssteven/ibert-roberta-base''': '''https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json''', '''kssteven/ibert-roberta-large''': '''https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json''', '''kssteven/ibert-roberta-large-mnli''': ( '''https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''ibert''' def __init__( self : Union[str, Any] , lowercase_ : Tuple=30522 , lowercase_ : Dict=768 , lowercase_ : Dict=12 , lowercase_ : Dict=12 , lowercase_ : Union[str, Any]=3072 , lowercase_ : Optional[Any]="gelu" , lowercase_ : Optional[Any]=0.1 , lowercase_ : int=0.1 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=2 , lowercase_ : Tuple=0.02 , lowercase_ : str=1e-1_2 , lowercase_ : Union[str, Any]=1 , lowercase_ : Optional[int]=0 , lowercase_ : List[Any]=2 , lowercase_ : Tuple="absolute" , lowercase_ : Any=False , lowercase_ : List[Any]="none" , **lowercase_ : Dict , ) -> Optional[Any]: """simple docstring""" super().__init__(pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = position_embedding_type _UpperCamelCase = quant_mode _UpperCamelCase = force_dequant class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": _UpperCamelCase = {0: "batch", 1: "choice", 2: "sequence"} else: _UpperCamelCase = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ])
702
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
0
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : str) -> List[str]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : str) -> Tuple: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = ScoreSdeVeScheduler() _UpperCamelCase = ScoreSdeVePipeline(unet=lowercase_ , scheduler=lowercase_) sde_ve.to(lowercase_) sde_ve.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = sde_ve(num_inference_steps=2 , output_type="numpy" , generator=lowercase_).images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = sde_ve(num_inference_steps=2 , output_type="numpy" , generator=lowercase_ , return_dict=lowercase_)[ 0 ] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = "google/ncsnpp-church-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = ScoreSdeVeScheduler.from_pretrained(lowercase_) _UpperCamelCase = ScoreSdeVePipeline(unet=lowercase_ , scheduler=lowercase_) sde_ve.to(lowercase_) sde_ve.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = sde_ve(num_inference_steps=10 , output_type="numpy" , generator=lowercase_).images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
703
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
82
0
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def lowerCAmelCase__ ( a__ ) ->bool: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = number while duplicate > 0: _UpperCamelCase , _UpperCamelCase = divmod(a__ , 10 ) fact_sum += factorial(a__ ) return fact_sum == number if __name__ == "__main__": print('''Program to check whether a number is a Krisnamurthy Number or not.''') lowerCamelCase__ = int(input('''Enter number: ''').strip()) print( F"{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number." )
704
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
0
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def lowerCAmelCase__ ( a__ , a__=None ) ->str: '''simple docstring''' _UpperCamelCase = None if token is not None: _UpperCamelCase = {"Accept": "application/vnd.github+json", "Authorization": f'Bearer {token}'} _UpperCamelCase = f'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' _UpperCamelCase = requests.get(a__ , headers=a__ ).json() _UpperCamelCase = {} try: job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) _UpperCamelCase = math.ceil((result["total_count"] - 100) / 100 ) for i in range(a__ ): _UpperCamelCase = requests.get(url + f'&page={i + 2}' , headers=a__ ).json() job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) return job_links except Exception: print(f'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def lowerCAmelCase__ ( a__ , a__=None ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = None if token is not None: _UpperCamelCase = {"Accept": "application/vnd.github+json", "Authorization": f'Bearer {token}'} _UpperCamelCase = f'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' _UpperCamelCase = requests.get(a__ , headers=a__ ).json() _UpperCamelCase = {} try: artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) _UpperCamelCase = math.ceil((result["total_count"] - 100) / 100 ) for i in range(a__ ): _UpperCamelCase = requests.get(url + f'&page={i + 2}' , headers=a__ ).json() artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) return artifacts except Exception: print(f'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = None if token is not None: _UpperCamelCase = {"Accept": "application/vnd.github+json", "Authorization": f'Bearer {token}'} _UpperCamelCase = requests.get(a__ , headers=a__ , allow_redirects=a__ ) _UpperCamelCase = result.headers["Location"] _UpperCamelCase = requests.get(a__ , allow_redirects=a__ ) _UpperCamelCase = os.path.join(a__ , f'{artifact_name}.zip' ) with open(a__ , "wb" ) as fp: fp.write(response.content ) def lowerCAmelCase__ ( a__ , a__=None ) ->str: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = None with zipfile.ZipFile(a__ ) as z: for filename in z.namelist(): if not os.path.isdir(a__ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(a__ ) as f: for line in f: _UpperCamelCase = line.decode("UTF-8" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs _UpperCamelCase = line[: line.index(": " )] _UpperCamelCase = line[line.index(": " ) + len(": " ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("FAILED " ): # `test` is the test method that failed _UpperCamelCase = line[len("FAILED " ) :] failed_tests.append(a__ ) elif filename == "job_name.txt": _UpperCamelCase = line if len(a__ ) != len(a__ ): raise ValueError( f'`errors` and `failed_tests` should have the same number of elements. Got {len(a__ )} for `errors` ' f'and {len(a__ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' " problem." ) _UpperCamelCase = None if job_name and job_links: _UpperCamelCase = job_links.get(a__ , a__ ) # A list with elements of the form (line of error, error, failed test) _UpperCamelCase = [x + [y] + [job_link] for x, y in zip(a__ , a__ )] return result def lowerCAmelCase__ ( a__ , a__=None ) ->Dict: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [os.path.join(a__ , a__ ) for p in os.listdir(a__ ) if p.endswith(".zip" )] for p in paths: errors.extend(get_errors_from_single_artifact(a__ , job_links=a__ ) ) return errors def lowerCAmelCase__ ( a__ , a__=None ) ->Tuple: '''simple docstring''' _UpperCamelCase = Counter() counter.update([x[1] for x in logs] ) _UpperCamelCase = counter.most_common() _UpperCamelCase = {} for error, count in counts: if error_filter is None or error not in error_filter: _UpperCamelCase = {"count": count, "failed_tests": [(x[2], x[0]) for x in logs if x[1] == error]} _UpperCamelCase = dict(sorted(r.items() , key=lambda a__ : item[1]["count"] , reverse=a__ ) ) return r def lowerCAmelCase__ ( a__ ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = test.split("::" )[0] if test.startswith("tests/models/" ): _UpperCamelCase = test.split("/" )[2] else: _UpperCamelCase = None return test def lowerCAmelCase__ ( a__ , a__=None ) ->int: '''simple docstring''' _UpperCamelCase = [(x[0], x[1], get_model(x[2] )) for x in logs] _UpperCamelCase = [x for x in logs if x[2] is not None] _UpperCamelCase = {x[2] for x in logs} _UpperCamelCase = {} for test in tests: _UpperCamelCase = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) _UpperCamelCase = counter.most_common() _UpperCamelCase = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} _UpperCamelCase = sum(error_counts.values() ) if n_errors > 0: _UpperCamelCase = {"count": n_errors, "errors": error_counts} _UpperCamelCase = dict(sorted(r.items() , key=lambda a__ : item[1]["count"] , reverse=a__ ) ) return r def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = "| no. | error | status |" _UpperCamelCase = "|-:|:-|:-|" _UpperCamelCase = [header, sep] for error in reduced_by_error: _UpperCamelCase = reduced_by_error[error]["count"] _UpperCamelCase = f'| {count} | {error[:100]} | |' lines.append(a__ ) return "\n".join(a__ ) def lowerCAmelCase__ ( a__ ) ->Tuple: '''simple docstring''' _UpperCamelCase = "| model | no. of errors | major error | count |" _UpperCamelCase = "|-:|-:|-:|-:|" _UpperCamelCase = [header, sep] for model in reduced_by_model: _UpperCamelCase = reduced_by_model[model]["count"] _UpperCamelCase , _UpperCamelCase = list(reduced_by_model[model]["errors"].items() )[0] _UpperCamelCase = f'| {model} | {count} | {error[:60]} | {_count} |' lines.append(a__ ) return "\n".join(a__ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') parser.add_argument( '''--output_dir''', type=str, required=True, help='''Where to store the downloaded artifacts and other result files.''', ) parser.add_argument('''--token''', default=None, type=str, help='''A token that has actions:read permission.''') lowerCamelCase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowerCamelCase__ = get_job_links(args.workflow_run_id, token=args.token) lowerCamelCase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowerCamelCase__ = k.find(''' / ''') lowerCamelCase__ = k[index + len(''' / ''') :] lowerCamelCase__ = v with open(os.path.join(args.output_dir, '''job_links.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowerCamelCase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, '''artifacts.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowerCamelCase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowerCamelCase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowerCamelCase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, '''errors.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowerCamelCase__ = reduce_by_error(errors) lowerCamelCase__ = reduce_by_model(errors) lowerCamelCase__ = make_github_table(reduced_by_error) lowerCamelCase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, '''reduced_by_error.txt'''), '''w''', encoding='''UTF-8''') as fp: fp.write(sa) with open(os.path.join(args.output_dir, '''reduced_by_model.txt'''), '''w''', encoding='''UTF-8''') as fp: fp.write(sa)
705
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
82
0
import os import sys import unittest lowerCamelCase__ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) lowerCamelCase__ = os.path.join('''tests''', '''models''', '''bert''', '''test_modeling_bert.py''') lowerCamelCase__ = os.path.join('''tests''', '''models''', '''blip''', '''test_modeling_blip.py''') class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> Dict: """simple docstring""" _UpperCamelCase = get_test_to_tester_mapping(lowercase_) _UpperCamelCase = get_test_to_tester_mapping(lowercase_) _UpperCamelCase = {"BertModelTest": "BertModelTester"} _UpperCamelCase = { "BlipModelTest": "BlipModelTester", "BlipTextImageModelTest": "BlipTextImageModelsModelTester", "BlipTextModelTest": "BlipTextModelTester", "BlipTextRetrievalModelTest": "BlipTextRetrievalModelTester", "BlipVQAModelTest": "BlipVQAModelTester", "BlipVisionModelTest": "BlipVisionModelTester", } self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_) self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_) def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = get_model_to_test_mapping(lowercase_) _UpperCamelCase = get_model_to_test_mapping(lowercase_) _UpperCamelCase = { "BertForMaskedLM": ["BertModelTest"], "BertForMultipleChoice": ["BertModelTest"], "BertForNextSentencePrediction": ["BertModelTest"], "BertForPreTraining": ["BertModelTest"], "BertForQuestionAnswering": ["BertModelTest"], "BertForSequenceClassification": ["BertModelTest"], "BertForTokenClassification": ["BertModelTest"], "BertLMHeadModel": ["BertModelTest"], "BertModel": ["BertModelTest"], } _UpperCamelCase = { "BlipForConditionalGeneration": ["BlipTextImageModelTest"], "BlipForImageTextRetrieval": ["BlipTextRetrievalModelTest"], "BlipForQuestionAnswering": ["BlipVQAModelTest"], "BlipModel": ["BlipModelTest"], "BlipTextModel": ["BlipTextModelTest"], "BlipVisionModel": ["BlipVisionModelTest"], } self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_) self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = get_model_to_tester_mapping(lowercase_) _UpperCamelCase = get_model_to_tester_mapping(lowercase_) _UpperCamelCase = { "BertForMaskedLM": ["BertModelTester"], "BertForMultipleChoice": ["BertModelTester"], "BertForNextSentencePrediction": ["BertModelTester"], "BertForPreTraining": ["BertModelTester"], "BertForQuestionAnswering": ["BertModelTester"], "BertForSequenceClassification": ["BertModelTester"], "BertForTokenClassification": ["BertModelTester"], "BertLMHeadModel": ["BertModelTester"], "BertModel": ["BertModelTester"], } _UpperCamelCase = { "BlipForConditionalGeneration": ["BlipTextImageModelsModelTester"], "BlipForImageTextRetrieval": ["BlipTextRetrievalModelTester"], "BlipForQuestionAnswering": ["BlipVQAModelTester"], "BlipModel": ["BlipModelTester"], "BlipTextModel": ["BlipTextModelTester"], "BlipVisionModel": ["BlipVisionModelTester"], } self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_) self.assertEqual(get_test_info.to_json(lowercase_) , lowercase_)
706
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
82
0
import sys import turtle def lowerCAmelCase__ ( a__ , a__ ) ->tuple[float, float]: '''simple docstring''' return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , ) ->None: '''simple docstring''' my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(a__ , get_mid(a__ , a__ ) , get_mid(a__ , a__ ) , depth - 1 ) triangle(a__ , get_mid(a__ , a__ ) , get_mid(a__ , a__ ) , depth - 1 ) triangle(a__ , get_mid(a__ , a__ ) , get_mid(a__ , a__ ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( '''Correct format for using this script: ''' '''python fractals.py <int:depth_for_fractal>''' ) lowerCamelCase__ = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor('''red''') lowerCamelCase__ = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
707
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
82
0
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
708
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
82
0
def lowerCAmelCase__ ( a__ ) ->list: '''simple docstring''' if len(a__ ) <= 1: return lst _UpperCamelCase = 1 while i < len(a__ ): if lst[i - 1] <= lst[i]: i += 1 else: _UpperCamelCase , _UpperCamelCase = lst[i], lst[i - 1] i -= 1 if i == 0: _UpperCamelCase = 1 return lst if __name__ == "__main__": lowerCamelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCamelCase__ = [int(item) for item in user_input.split(''',''')] print(gnome_sort(unsorted))
709
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , *lowercase_ : List[str] , **lowercase_ : Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
82
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCamelCase__ = {'''configuration_unispeech''': ['''UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''UniSpeechConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST''', '''UniSpeechForCTC''', '''UniSpeechForPreTraining''', '''UniSpeechForSequenceClassification''', '''UniSpeechModel''', '''UniSpeechPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
710
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
82
0
'''simple docstring''' import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging lowerCamelCase__ = logging.get_logger(__name__) logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ ) ->str: '''simple docstring''' if "xprophetnet" in prophetnet_checkpoint_path: _UpperCamelCase = XLMProphetNetForConditionalGenerationOld.from_pretrained(a__ ) _UpperCamelCase , _UpperCamelCase = XLMProphetNetForConditionalGeneration.from_pretrained( a__ , output_loading_info=a__ ) else: _UpperCamelCase = ProphetNetForConditionalGenerationOld.from_pretrained(a__ ) _UpperCamelCase , _UpperCamelCase = ProphetNetForConditionalGeneration.from_pretrained( a__ , output_loading_info=a__ ) _UpperCamelCase = ["key_proj", "value_proj", "query_proj"] _UpperCamelCase = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: _UpperCamelCase = key.split("." ) if attributes[0] == "lm_head": _UpperCamelCase = prophet _UpperCamelCase = prophet_old else: _UpperCamelCase = prophet.prophetnet _UpperCamelCase = prophet_old.model _UpperCamelCase = False for attribute in attributes: if attribute in mapping: _UpperCamelCase = mapping[attribute] if not hasattr(a__ , a__ ) and len(a__ ) > 0: _UpperCamelCase = attribute elif hasattr(a__ , a__ ): _UpperCamelCase = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _UpperCamelCase = old_model.weight logger.info(f'{attribute} is initialized.' ) _UpperCamelCase = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _UpperCamelCase = old_model.bias logger.info(f'{attribute} is initialized' ) _UpperCamelCase = True break elif attribute in special_keys and hasattr(a__ , "in_proj_weight" ): _UpperCamelCase = old_model.in_proj_weight.shape[0] // 3 _UpperCamelCase = getattr(a__ , a__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _UpperCamelCase = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _UpperCamelCase = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _UpperCamelCase = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _UpperCamelCase = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _UpperCamelCase = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _UpperCamelCase = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _UpperCamelCase = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." _UpperCamelCase = nn.Parameter(old_model.embed_positions.weight[:512, :] ) _UpperCamelCase = True break if attribute.isdigit(): _UpperCamelCase = model[int(a__ )] _UpperCamelCase = old_model[int(a__ )] else: _UpperCamelCase = getattr(a__ , a__ ) if old_attribute == "": _UpperCamelCase = old_model else: if not hasattr(a__ , a__ ): raise ValueError(f'{old_model} does not have {old_attribute}' ) _UpperCamelCase = getattr(a__ , a__ ) if not is_key_init: raise ValueError(f'{key} was not correctly initialized!' ) print(f'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(a__ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
711
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
82
0
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = CodeGenTokenizer __A = CodeGenTokenizerFast __A = True __A = {'''add_prefix_space''': True} __A = False def __UpperCAmelCase ( self : Optional[int]) -> Optional[Any]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCamelCase = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] _UpperCamelCase = dict(zip(lowercase_ , range(len(lowercase_)))) _UpperCamelCase = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] _UpperCamelCase = {"unk_token": "<unk>"} _UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) _UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(lowercase_) + "\n") with open(self.merges_file , "w" , encoding="utf-8") as fp: fp.write("\n".join(lowercase_)) def __UpperCAmelCase ( self : Tuple , **lowercase_ : Any) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **lowercase_) def __UpperCAmelCase ( self : Tuple , **lowercase_ : Tuple) -> Optional[Any]: """simple docstring""" kwargs.update(self.special_tokens_map) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_) def __UpperCAmelCase ( self : Any , lowercase_ : List[str]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = "lower newer" _UpperCamelCase = "lower newer" return input_text, output_text def __UpperCAmelCase ( self : int) -> Optional[Any]: """simple docstring""" _UpperCamelCase = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) _UpperCamelCase = "lower newer" _UpperCamelCase = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] _UpperCamelCase = tokenizer.tokenize(lowercase_ , add_prefix_space=lowercase_) self.assertListEqual(lowercase_ , lowercase_) _UpperCamelCase = tokens + [tokenizer.unk_token] _UpperCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase_) , lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Any: """simple docstring""" if not self.test_rust_tokenizer: return _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=lowercase_) _UpperCamelCase = "lower newer" # Testing tokenization _UpperCamelCase = tokenizer.tokenize(lowercase_ , add_prefix_space=lowercase_) _UpperCamelCase = rust_tokenizer.tokenize(lowercase_) self.assertListEqual(lowercase_ , lowercase_) # Testing conversion to ids without special tokens _UpperCamelCase = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ , add_prefix_space=lowercase_) _UpperCamelCase = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) self.assertListEqual(lowercase_ , lowercase_) # Testing conversion to ids with special tokens _UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=lowercase_) _UpperCamelCase = tokenizer.encode(lowercase_ , add_prefix_space=lowercase_) _UpperCamelCase = rust_tokenizer.encode(lowercase_) self.assertListEqual(lowercase_ , lowercase_) # Testing the unknown token _UpperCamelCase = tokens + [rust_tokenizer.unk_token] _UpperCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(lowercase_) , lowercase_) def __UpperCAmelCase ( self : str , *lowercase_ : int , **lowercase_ : List[Any]) -> Tuple: """simple docstring""" pass def __UpperCAmelCase ( self : Tuple , lowercase_ : Optional[int]=15) -> List[Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_) # Simple input _UpperCamelCase = "This is a simple input" _UpperCamelCase = ["This is a simple input 1", "This is a simple input 2"] _UpperCamelCase = ("This is a simple input", "This is a pair") _UpperCamelCase = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length") # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length") # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length") # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length") # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def __UpperCAmelCase ( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>") # Simple input _UpperCamelCase = "This is a simple input" _UpperCamelCase = ["This is a simple input looooooooong", "This is a simple input"] _UpperCamelCase = ("This is a simple input", "This is a pair") _UpperCamelCase = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] _UpperCamelCase = tokenizer.pad_token_id _UpperCamelCase = tokenizer(lowercase_ , padding="max_length" , max_length=30 , return_tensors="np") _UpperCamelCase = tokenizer(lowercase_ , padding=lowercase_ , truncate=lowercase_ , return_tensors="np") _UpperCamelCase = tokenizer(*lowercase_ , padding="max_length" , max_length=60 , return_tensors="np") _UpperCamelCase = tokenizer(lowercase_ , padding=lowercase_ , truncate=lowercase_ , return_tensors="np") # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 30) self.assertTrue(pad_token_id in out_s["input_ids"]) self.assertTrue(0 in out_s["attention_mask"]) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 33) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0]) self.assertFalse(0 in out_sa["attention_mask"][0]) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1]) self.assertTrue(0 in out_sa["attention_mask"][1]) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 60) self.assertTrue(pad_token_id in out_p["input_ids"]) self.assertTrue(0 in out_p["attention_mask"]) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 52) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0]) self.assertFalse(0 in out_pa["attention_mask"][0]) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1]) self.assertTrue(0 in out_pa["attention_mask"][1]) def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" _UpperCamelCase = "$$$" _UpperCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=lowercase_ , add_bos_token=lowercase_) _UpperCamelCase = "This is a simple input" _UpperCamelCase = ["This is a simple input 1", "This is a simple input 2"] _UpperCamelCase = tokenizer.bos_token_id _UpperCamelCase = tokenizer(lowercase_) _UpperCamelCase = tokenizer(lowercase_) self.assertEqual(out_s.input_ids[0] , lowercase_) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids)) _UpperCamelCase = tokenizer.decode(out_s.input_ids) _UpperCamelCase = tokenizer.batch_decode(out_sa.input_ids) self.assertEqual(decode_s.split()[0] , lowercase_) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa)) @slow def __UpperCAmelCase ( self : Optional[Any]) -> str: """simple docstring""" _UpperCamelCase = CodeGenTokenizer.from_pretrained("Salesforce/codegen-350M-mono") _UpperCamelCase = "\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#" _UpperCamelCase = "\nif len_a > len_b: result = a\nelse: result = b" _UpperCamelCase = tokenizer.encode(lowercase_) _UpperCamelCase = ["^#", re.escape("<|endoftext|>"), "^'''", "^\"\"\"", "\n\n\n"] _UpperCamelCase = tokenizer.decode(lowercase_ , truncate_before_pattern=lowercase_) self.assertEqual(lowercase_ , lowercase_) def __UpperCAmelCase ( self : List[str]) -> Optional[int]: """simple docstring""" pass
712
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
82
0
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" _UpperCamelCase = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowercase_ , "hidden_sizes")) self.parent.assertTrue(hasattr(lowercase_ , "num_attention_heads")) self.parent.assertTrue(hasattr(lowercase_ , "num_encoder_blocks")) class _UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Tuple=13 , lowercase_ : List[Any]=64 , lowercase_ : List[Any]=3 , lowercase_ : Optional[Any]=4 , lowercase_ : Dict=[2, 2, 2, 2] , lowercase_ : Optional[Any]=[8, 4, 2, 1] , lowercase_ : str=[16, 32, 64, 128] , lowercase_ : Optional[Any]=[1, 4, 8, 16] , lowercase_ : Tuple=[1, 2, 4, 8] , lowercase_ : List[str]=True , lowercase_ : int=True , lowercase_ : Optional[int]="gelu" , lowercase_ : Dict=0.1 , lowercase_ : List[Any]=0.1 , lowercase_ : str=0.02 , lowercase_ : Optional[int]=3 , lowercase_ : List[str]=None , ) -> List[Any]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = num_channels _UpperCamelCase = num_encoder_blocks _UpperCamelCase = sr_ratios _UpperCamelCase = depths _UpperCamelCase = hidden_sizes _UpperCamelCase = downsampling_rates _UpperCamelCase = num_attention_heads _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = scope def __UpperCAmelCase ( self : Any) -> Dict: """simple docstring""" _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels) _UpperCamelCase = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self : Any , lowercase_ : Dict , lowercase_ : List[str] , lowercase_ : str) -> List[str]: """simple docstring""" _UpperCamelCase = SegformerModel(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_) _UpperCamelCase = _UpperCamelCase = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width)) def __UpperCAmelCase ( self : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = SegformerForSemanticSegmentation(lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) _UpperCamelCase = model(lowercase_ , labels=lowercase_) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) self.parent.assertGreater(result.loss , 0.0) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : int) -> str: """simple docstring""" _UpperCamelCase = 1 _UpperCamelCase = SegformerForSemanticSegmentation(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size)).to(lowercase_) _UpperCamelCase = model(lowercase_ , labels=lowercase_) self.parent.assertGreater(result.loss , 0.0) def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = config_and_inputs _UpperCamelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( lowerCAmelCase, lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __A = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __A = True __A = False __A = False __A = False def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" _UpperCamelCase = SegformerModelTester(self) _UpperCamelCase = SegformerConfigTester(self , config_class=lowercase_) def __UpperCAmelCase ( self : Tuple) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : str) -> Tuple: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_) def __UpperCAmelCase ( self : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*lowercase_) def __UpperCAmelCase ( self : Any) -> Any: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*lowercase_) @unittest.skip("SegFormer does not use inputs_embeds") def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods") def __UpperCAmelCase ( self : str) -> List[str]: """simple docstring""" pass def __UpperCAmelCase ( self : Tuple) -> str: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> List[str]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True for model_class in self.all_model_classes: _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.attentions _UpperCamelCase = sum(self.model_tester.depths) self.assertEqual(len(lowercase_) , lowercase_) # check that output_attentions also work using config del inputs_dict["output_attentions"] _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.attentions self.assertEqual(len(lowercase_) , lowercase_) # verify the first attentions (first block, first layer) _UpperCamelCase = (self.model_tester.image_size // 4) ** 2 _UpperCamelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) _UpperCamelCase = (self.model_tester.image_size // 32) ** 2 _UpperCamelCase = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:]) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) _UpperCamelCase = len(lowercase_) # Check attention is always last and order is fine _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) self.assertEqual(out_len + 1 , len(lowercase_)) _UpperCamelCase = outputs.attentions self.assertEqual(len(lowercase_) , lowercase_) # verify the first attentions (first block, first layer) _UpperCamelCase = (self.model_tester.image_size // 4) ** 2 _UpperCamelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" def check_hidden_states_output(lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Optional[int]): _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_)) _UpperCamelCase = outputs.hidden_states _UpperCamelCase = self.model_tester.num_encoder_blocks self.assertEqual(len(lowercase_) , lowercase_) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:]) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_) def __UpperCAmelCase ( self : Any) -> Optional[Any]: """simple docstring""" if not self.model_tester.is_training: return _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True for model_class in self.all_model_classes: if model_class in get_values(lowercase_): continue _UpperCamelCase = model_class(lowercase_) model.to(lowercase_) model.train() _UpperCamelCase = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_) _UpperCamelCase = model(**lowercase_).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests.") def __UpperCAmelCase ( self : Any) -> Union[str, Any]: """simple docstring""" pass @slow def __UpperCAmelCase ( self : Dict) -> Optional[Any]: """simple docstring""" for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = SegformerModel.from_pretrained(lowercase_) self.assertIsNotNone(lowercase_) def lowerCAmelCase__ ( ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self : List[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=lowercase_ , align=lowercase_ , do_random_crop=lowercase_) _UpperCamelCase = SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512").to( lowercase_) _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=lowercase_ , return_tensors="pt") _UpperCamelCase = encoded_inputs.pixel_values.to(lowercase_) with torch.no_grad(): _UpperCamelCase = model(lowercase_) _UpperCamelCase = torch.Size((1, model.config.num_labels, 128, 128)) self.assertEqual(outputs.logits.shape , lowercase_) _UpperCamelCase = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ]).to(lowercase_) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowercase_ , atol=1e-4)) @slow def __UpperCAmelCase ( self : Tuple) -> List[str]: """simple docstring""" _UpperCamelCase = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=lowercase_ , align=lowercase_ , do_random_crop=lowercase_) _UpperCamelCase = SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024").to(lowercase_) _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=lowercase_ , return_tensors="pt") _UpperCamelCase = encoded_inputs.pixel_values.to(lowercase_) with torch.no_grad(): _UpperCamelCase = model(lowercase_) _UpperCamelCase = torch.Size((1, model.config.num_labels, 128, 128)) self.assertEqual(outputs.logits.shape , lowercase_) _UpperCamelCase = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ]).to(lowercase_) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowercase_ , atol=1e-1)) @slow def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=lowercase_ , align=lowercase_ , do_random_crop=lowercase_) _UpperCamelCase = SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512").to( lowercase_) _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=lowercase_ , return_tensors="pt") _UpperCamelCase = encoded_inputs.pixel_values.to(lowercase_) with torch.no_grad(): _UpperCamelCase = model(lowercase_) _UpperCamelCase = outputs.logits.detach().cpu() _UpperCamelCase = image_processor.post_process_semantic_segmentation(outputs=lowercase_ , target_sizes=[(500, 300)]) _UpperCamelCase = torch.Size((500, 300)) self.assertEqual(segmentation[0].shape , lowercase_) _UpperCamelCase = image_processor.post_process_semantic_segmentation(outputs=lowercase_) _UpperCamelCase = torch.Size((128, 128)) self.assertEqual(segmentation[0].shape , lowercase_)
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
from bisect import bisect from itertools import accumulate def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Tuple: '''simple docstring''' _UpperCamelCase = sorted(zip(a__ , a__ ) , key=lambda a__ : x[0] / x[1] , reverse=a__ ) _UpperCamelCase , _UpperCamelCase = [i[0] for i in r], [i[1] for i in r] _UpperCamelCase = list(accumulate(a__ ) ) _UpperCamelCase = bisect(a__ , a__ ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask lowerCamelCase__ = logging.getLogger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''token-classification''' def __init__( self : str , lowercase_ : Any) -> str: """simple docstring""" if type(lowercase_) == dict: _UpperCamelCase = Namespace(**lowercase_) _UpperCamelCase = import_module("tasks") try: _UpperCamelCase = getattr(lowercase_ , hparams.task_type) _UpperCamelCase = token_classification_task_clazz() except AttributeError: raise ValueError( f'Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ' f'Available tasks classes are: {TokenClassificationTask.__subclasses__()}') _UpperCamelCase = self.token_classification_task.get_labels(hparams.labels) _UpperCamelCase = CrossEntropyLoss().ignore_index super().__init__(lowercase_ , len(self.labels) , self.mode) def __UpperCAmelCase ( self : List[Any] , **lowercase_ : Optional[int]) -> Dict: """simple docstring""" return self.model(**lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Union[str, Any] , lowercase_ : List[str]) -> int: """simple docstring""" _UpperCamelCase = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": _UpperCamelCase = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids _UpperCamelCase = self(**lowercase_) _UpperCamelCase = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def __UpperCAmelCase ( self : List[str]) -> Dict: """simple docstring""" _UpperCamelCase = self.hparams for mode in ["train", "dev", "test"]: _UpperCamelCase = self._feature_file(lowercase_) if os.path.exists(lowercase_) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , lowercase_) _UpperCamelCase = torch.load(lowercase_) else: logger.info("Creating features from dataset file at %s" , args.data_dir) _UpperCamelCase = self.token_classification_task.read_examples_from_file(args.data_dir , lowercase_) _UpperCamelCase = self.token_classification_task.convert_examples_to_features( lowercase_ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"]) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=lowercase_ , pad_on_left=bool(self.config.model_type in ["xlnet"]) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , lowercase_) torch.save(lowercase_ , lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , lowercase_ : int , lowercase_ : bool = False) -> DataLoader: """simple docstring""" _UpperCamelCase = self._feature_file(lowercase_) logger.info("Loading features from cached file %s" , lowercase_) _UpperCamelCase = torch.load(lowercase_) _UpperCamelCase = torch.tensor([f.input_ids for f in features] , dtype=torch.long) _UpperCamelCase = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) if features[0].token_type_ids is not None: _UpperCamelCase = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) else: _UpperCamelCase = torch.tensor([0 for f in features] , dtype=torch.long) # HACK(we will not use this anymore soon) _UpperCamelCase = torch.tensor([f.label_ids for f in features] , dtype=torch.long) return DataLoader( TensorDataset(lowercase_ , lowercase_ , lowercase_ , lowercase_) , batch_size=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Tuple , lowercase_ : str) -> Dict: """simple docstring""" """Compute validation""" "" _UpperCamelCase = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": _UpperCamelCase = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids _UpperCamelCase = self(**lowercase_) _UpperCamelCase , _UpperCamelCase = outputs[:2] _UpperCamelCase = logits.detach().cpu().numpy() _UpperCamelCase = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = torch.stack([x["val_loss"] for x in outputs]).mean() _UpperCamelCase = np.concatenate([x["pred"] for x in outputs] , axis=0) _UpperCamelCase = np.argmax(lowercase_ , axis=2) _UpperCamelCase = np.concatenate([x["target"] for x in outputs] , axis=0) _UpperCamelCase = dict(enumerate(self.labels)) _UpperCamelCase = [[] for _ in range(out_label_ids.shape[0])] _UpperCamelCase = [[] for _ in range(out_label_ids.shape[0])] for i in range(out_label_ids.shape[0]): for j in range(out_label_ids.shape[1]): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]]) preds_list[i].append(label_map[preds[i][j]]) _UpperCamelCase = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(lowercase_ , lowercase_), "precision": precision_score(lowercase_ , lowercase_), "recall": recall_score(lowercase_ , lowercase_), "f1": fa_score(lowercase_ , lowercase_), } _UpperCamelCase = dict(results.items()) _UpperCamelCase = results return ret, preds_list, out_label_list def __UpperCAmelCase ( self : int , lowercase_ : Any) -> Optional[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self._eval_end(lowercase_) _UpperCamelCase = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> str: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self._eval_end(lowercase_) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 _UpperCamelCase = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def __UpperCAmelCase ( lowercase_ : List[str] , lowercase_ : int) -> Any: """simple docstring""" BaseTransformer.add_model_specific_args(lowercase_ , lowercase_) parser.add_argument( "--task_type" , default="NER" , type=lowercase_ , help="Task type to fine tune in training (e.g. NER, POS, etc)") parser.add_argument( "--max_seq_length" , default=128 , type=lowercase_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=lowercase_ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=lowercase_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets") return parser if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) lowerCamelCase__ = NERTransformer.add_model_specific_args(parser, os.getcwd()) lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = NERTransformer(args) lowerCamelCase__ = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 lowerCamelCase__ = sorted(glob.glob(os.path.join(args.output_dir, '''checkpoint-epoch=*.ckpt'''), recursive=True)) lowerCamelCase__ = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) lowerCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path" , type=a__ , default="data/dump.txt" , help="The path to the data." ) parser.add_argument("--tokenizer_type" , type=a__ , default="bert" , choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name" , type=a__ , default="bert-base-uncased" , help="The tokenizer to use." ) parser.add_argument("--dump_file" , type=a__ , default="data/dump" , help="The dump file prefix." ) _UpperCamelCase = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase = tokenizer.special_tokens_map["cls_token"] # `[CLS]` _UpperCamelCase = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase = tokenizer.special_tokens_map["cls_token"] # `<s>` _UpperCamelCase = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` _UpperCamelCase = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , "r" , encoding="utf8" ) as fp: _UpperCamelCase = fp.readlines() logger.info("Start encoding" ) logger.info(f'{len(a__ )} examples to process.' ) _UpperCamelCase = [] _UpperCamelCase = 0 _UpperCamelCase = 10_000 _UpperCamelCase = time.time() for text in data: _UpperCamelCase = f'{bos} {text.strip()} {sep}' _UpperCamelCase = tokenizer.encode(a__ , add_special_tokens=a__ ) rslt.append(a__ ) iter += 1 if iter % interval == 0: _UpperCamelCase = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase = time.time() logger.info("Finished binarization" ) logger.info(f'{len(a__ )} examples processed.' ) _UpperCamelCase = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase = tokenizer.vocab_size if vocab_size < (1 << 16): _UpperCamelCase = [np.uintaa(a__ ) for d in rslt] else: _UpperCamelCase = [np.intaa(a__ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(a__ , "wb" ) as handle: pickle.dump(rslt_ , a__ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = '''RegNetConfig''' # Base docstring lowerCamelCase__ = '''facebook/regnet-y-040''' lowerCamelCase__ = [1, 1088, 7, 7] # Image classification docstring lowerCamelCase__ = '''facebook/regnet-y-040''' lowerCamelCase__ = '''tabby, tabby cat''' lowerCamelCase__ = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : int , lowercase_ : int , lowercase_ : int = 3 , lowercase_ : int = 1 , lowercase_ : int = 1 , lowercase_ : Optional[str] = "relu" , ) -> Tuple: """simple docstring""" super().__init__() _UpperCamelCase = nn.Convad( lowercase_ , lowercase_ , kernel_size=lowercase_ , stride=lowercase_ , padding=kernel_size // 2 , groups=lowercase_ , bias=lowercase_ , ) _UpperCamelCase = nn.BatchNormad(lowercase_) _UpperCamelCase = ACTaFN[activation] if activation is not None else nn.Identity() def __UpperCAmelCase ( self : str , lowercase_ : List[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.convolution(lowercase_) _UpperCamelCase = self.normalization(lowercase_) _UpperCamelCase = self.activation(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : List[Any] , lowercase_ : RegNetConfig) -> str: """simple docstring""" super().__init__() _UpperCamelCase = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act) _UpperCamelCase = config.num_channels def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any) -> Optional[Any]: """simple docstring""" _UpperCamelCase = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration.") _UpperCamelCase = self.embedder(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : int , lowercase_ : int , lowercase_ : int = 2) -> Tuple: """simple docstring""" super().__init__() _UpperCamelCase = nn.Convad(lowercase_ , lowercase_ , kernel_size=1 , stride=lowercase_ , bias=lowercase_) _UpperCamelCase = nn.BatchNormad(lowercase_) def __UpperCAmelCase ( self : str , lowercase_ : Tensor) -> Tensor: """simple docstring""" _UpperCamelCase = self.convolution(lowercase_) _UpperCamelCase = self.normalization(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Dict , lowercase_ : int , lowercase_ : int) -> List[Any]: """simple docstring""" super().__init__() _UpperCamelCase = nn.AdaptiveAvgPoolad((1, 1)) _UpperCamelCase = nn.Sequential( nn.Convad(lowercase_ , lowercase_ , kernel_size=1) , nn.ReLU() , nn.Convad(lowercase_ , lowercase_ , kernel_size=1) , nn.Sigmoid() , ) def __UpperCAmelCase ( self : int , lowercase_ : Union[str, Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.pooler(lowercase_) _UpperCamelCase = self.attention(lowercase_) _UpperCamelCase = hidden_state * attention return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : RegNetConfig , lowercase_ : int , lowercase_ : int , lowercase_ : int = 1) -> Dict: """simple docstring""" super().__init__() _UpperCamelCase = in_channels != out_channels or stride != 1 _UpperCamelCase = max(1 , out_channels // config.groups_width) _UpperCamelCase = ( RegNetShortCut(lowercase_ , lowercase_ , stride=lowercase_) if should_apply_shortcut else nn.Identity() ) _UpperCamelCase = nn.Sequential( RegNetConvLayer(lowercase_ , lowercase_ , kernel_size=1 , activation=config.hidden_act) , RegNetConvLayer(lowercase_ , lowercase_ , stride=lowercase_ , groups=lowercase_ , activation=config.hidden_act) , RegNetConvLayer(lowercase_ , lowercase_ , kernel_size=1 , activation=lowercase_) , ) _UpperCamelCase = ACTaFN[config.hidden_act] def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> Optional[Any]: """simple docstring""" _UpperCamelCase = hidden_state _UpperCamelCase = self.layer(lowercase_) _UpperCamelCase = self.shortcut(lowercase_) hidden_state += residual _UpperCamelCase = self.activation(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Dict , lowercase_ : RegNetConfig , lowercase_ : int , lowercase_ : int , lowercase_ : int = 1) -> List[str]: """simple docstring""" super().__init__() _UpperCamelCase = in_channels != out_channels or stride != 1 _UpperCamelCase = max(1 , out_channels // config.groups_width) _UpperCamelCase = ( RegNetShortCut(lowercase_ , lowercase_ , stride=lowercase_) if should_apply_shortcut else nn.Identity() ) _UpperCamelCase = nn.Sequential( RegNetConvLayer(lowercase_ , lowercase_ , kernel_size=1 , activation=config.hidden_act) , RegNetConvLayer(lowercase_ , lowercase_ , stride=lowercase_ , groups=lowercase_ , activation=config.hidden_act) , RegNetSELayer(lowercase_ , reduced_channels=int(round(in_channels / 4))) , RegNetConvLayer(lowercase_ , lowercase_ , kernel_size=1 , activation=lowercase_) , ) _UpperCamelCase = ACTaFN[config.hidden_act] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Tuple) -> Dict: """simple docstring""" _UpperCamelCase = hidden_state _UpperCamelCase = self.layer(lowercase_) _UpperCamelCase = self.shortcut(lowercase_) hidden_state += residual _UpperCamelCase = self.activation(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : Any , lowercase_ : RegNetConfig , lowercase_ : int , lowercase_ : int , lowercase_ : int = 2 , lowercase_ : int = 2 , ) -> int: """simple docstring""" super().__init__() _UpperCamelCase = RegNetXLayer if config.layer_type == "x" else RegNetYLayer _UpperCamelCase = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( lowercase_ , lowercase_ , lowercase_ , stride=lowercase_ , ) , *[layer(lowercase_ , lowercase_ , lowercase_) for _ in range(depth - 1)] , ) def __UpperCAmelCase ( self : Dict , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = self.layers(lowercase_) return hidden_state class _UpperCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self : List[Any] , lowercase_ : RegNetConfig) -> str: """simple docstring""" super().__init__() _UpperCamelCase = nn.ModuleList([]) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( lowercase_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , )) _UpperCamelCase = zip(config.hidden_sizes , config.hidden_sizes[1:]) for (in_channels, out_channels), depth in zip(lowercase_ , config.depths[1:]): self.stages.append(RegNetStage(lowercase_ , lowercase_ , lowercase_ , depth=lowercase_)) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Tensor , lowercase_ : bool = False , lowercase_ : bool = True) -> BaseModelOutputWithNoAttention: """simple docstring""" _UpperCamelCase = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCamelCase = hidden_states + (hidden_state,) _UpperCamelCase = stage_module(lowercase_) if output_hidden_states: _UpperCamelCase = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None) return BaseModelOutputWithNoAttention(last_hidden_state=lowercase_ , hidden_states=lowercase_) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = RegNetConfig __A = '''regnet''' __A = '''pixel_values''' __A = True def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple) -> List[str]: """simple docstring""" if isinstance(lowercase_ , nn.Convad): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu") elif isinstance(lowercase_ , (nn.BatchNormad, nn.GroupNorm)): nn.init.constant_(module.weight , 1) nn.init.constant_(module.bias , 0) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Dict , lowercase_ : Union[str, Any]=False) -> int: """simple docstring""" if isinstance(lowercase_ , lowercase_): _UpperCamelCase = value lowerCamelCase__ = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' lowerCamelCase__ = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''', lowerCAmelCase, ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Tuple , lowercase_ : Optional[int]) -> Tuple: """simple docstring""" super().__init__(lowercase_) _UpperCamelCase = config _UpperCamelCase = RegNetEmbeddings(lowercase_) _UpperCamelCase = RegNetEncoder(lowercase_) _UpperCamelCase = nn.AdaptiveAvgPoolad((1, 1)) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowercase_ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __UpperCAmelCase ( self : Dict , lowercase_ : Tensor , lowercase_ : Optional[bool] = None , lowercase_ : Optional[bool] = None) -> BaseModelOutputWithPoolingAndNoAttention: """simple docstring""" _UpperCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCamelCase = return_dict if return_dict is not None else self.config.use_return_dict _UpperCamelCase = self.embedder(lowercase_) _UpperCamelCase = self.encoder( lowercase_ , output_hidden_states=lowercase_ , return_dict=lowercase_) _UpperCamelCase = encoder_outputs[0] _UpperCamelCase = self.pooler(lowercase_) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowercase_ , pooler_output=lowercase_ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( ''' RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''', lowerCAmelCase, ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : Any) -> Any: """simple docstring""" super().__init__(lowercase_) _UpperCamelCase = config.num_labels _UpperCamelCase = RegNetModel(lowercase_) # classification head _UpperCamelCase = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowercase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Optional[torch.FloatTensor] = None , lowercase_ : Optional[torch.LongTensor] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: """simple docstring""" _UpperCamelCase = return_dict if return_dict is not None else self.config.use_return_dict _UpperCamelCase = self.regnet(lowercase_ , output_hidden_states=lowercase_ , return_dict=lowercase_) _UpperCamelCase = outputs.pooler_output if return_dict else outputs[1] _UpperCamelCase = self.classifier(lowercase_) _UpperCamelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _UpperCamelCase = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _UpperCamelCase = "single_label_classification" else: _UpperCamelCase = "multi_label_classification" if self.config.problem_type == "regression": _UpperCamelCase = MSELoss() if self.num_labels == 1: _UpperCamelCase = loss_fct(logits.squeeze() , labels.squeeze()) else: _UpperCamelCase = loss_fct(lowercase_ , lowercase_) elif self.config.problem_type == "single_label_classification": _UpperCamelCase = CrossEntropyLoss() _UpperCamelCase = loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) elif self.config.problem_type == "multi_label_classification": _UpperCamelCase = BCEWithLogitsLoss() _UpperCamelCase = loss_fct(lowercase_ , lowercase_) if not return_dict: _UpperCamelCase = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowercase_ , logits=lowercase_ , hidden_states=outputs.hidden_states)
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) lowerCamelCase__ = logging.getLogger(__name__) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser( description='''Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)''' ) parser.add_argument( '''--data_file''', type=str, default='''data/dump.bert-base-uncased.pickle''', help='''The binarized dataset.''' ) parser.add_argument( '''--token_counts_dump''', type=str, default='''data/token_counts.bert-base-uncased.pickle''', help='''The dump file.''' ) parser.add_argument('''--vocab_size''', default=3_0522, type=int) lowerCamelCase__ = parser.parse_args() logger.info(F"Loading data from {args.data_file}") with open(args.data_file, '''rb''') as fp: lowerCamelCase__ = pickle.load(fp) logger.info('''Counting occurrences for MLM.''') lowerCamelCase__ = Counter() for tk_ids in data: counter.update(tk_ids) lowerCamelCase__ = [0] * args.vocab_size for k, v in counter.items(): lowerCamelCase__ = v logger.info(F"Dump to {args.token_counts_dump}") with open(args.token_counts_dump, '''wb''') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
from __future__ import annotations lowerCamelCase__ = [True] * 100_0001 lowerCamelCase__ = 2 while i * i <= 100_0000: if seive[i]: for j in range(i * i, 100_0001, i): lowerCamelCase__ = False i += 1 def lowerCAmelCase__ ( a__ ) ->bool: '''simple docstring''' return seive[n] def lowerCAmelCase__ ( a__ ) ->bool: '''simple docstring''' return any(digit in "02468" for digit in str(a__ ) ) def lowerCAmelCase__ ( a__ = 1_000_000 ) ->list[int]: '''simple docstring''' _UpperCamelCase = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(a__ ) and not contains_an_even_digit(a__ ): _UpperCamelCase = str(a__ ) _UpperCamelCase = [int(str_num[j:] + str_num[:j] ) for j in range(len(a__ ) )] if all(is_prime(a__ ) for i in list_nums ): result.append(a__ ) return result def lowerCAmelCase__ ( ) ->int: '''simple docstring''' return len(find_circular_primes() ) if __name__ == "__main__": print(F"{len(find_circular_primes()) = }")
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase__ = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongformerConfig''', '''LongformerOnnxConfig''', ], '''tokenization_longformer''': ['''LongformerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['''LongformerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongformerForMaskedLM''', '''LongformerForMultipleChoice''', '''LongformerForQuestionAnswering''', '''LongformerForSequenceClassification''', '''LongformerForTokenClassification''', '''LongformerModel''', '''LongformerPreTrainedModel''', '''LongformerSelfAttention''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLongformerForMaskedLM''', '''TFLongformerForMultipleChoice''', '''TFLongformerForQuestionAnswering''', '''TFLongformerForSequenceClassification''', '''TFLongformerForTokenClassification''', '''TFLongformerModel''', '''TFLongformerPreTrainedModel''', '''TFLongformerSelfAttention''', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = 9, 14 # noqa: F841 _UpperCamelCase = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] _UpperCamelCase = defaultdict(a__ ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) _UpperCamelCase = mst(a__ ) _UpperCamelCase = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: _UpperCamelCase = tuple(answer[:2] ) _UpperCamelCase = tuple(edge[::-1] ) assert edge in result or reverse in result
700
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
82
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( a__ , a__=False ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def lowerCAmelCase__ ( a__ , a__ , a__=False ) ->List[Any]: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = "" else: _UpperCamelCase = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) _UpperCamelCase = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCAmelCase__ ( a__ ) ->Tuple: '''simple docstring''' _UpperCamelCase = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(a__ , a__ ) def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = dct.pop(a__ ) _UpperCamelCase = val def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCamelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( a__ , a__ , a__=True ) ->Dict: '''simple docstring''' _UpperCamelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCamelCase = 8 # set labels if required if not base_model: _UpperCamelCase = 1_000 _UpperCamelCase = "huggingface/label-files" _UpperCamelCase = "imagenet-1k-id2label.json" _UpperCamelCase = json.load(open(hf_hub_download(a__ , a__ , repo_type="dataset" ) , "r" ) ) _UpperCamelCase = {int(a__ ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCamelCase = 384 _UpperCamelCase = 1_536 _UpperCamelCase = 12 _UpperCamelCase = 6 # load original model from torch hub _UpperCamelCase = torch.hub.load("facebookresearch/dino:main" , a__ ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = original_model.state_dict() if base_model: remove_classification_head_(a__ ) _UpperCamelCase = create_rename_keys(a__ , base_model=a__ ) for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) read_in_q_k_v(a__ , a__ , a__ ) # load HuggingFace model if base_model: _UpperCamelCase = ViTModel(a__ , add_pooling_layer=a__ ).eval() else: _UpperCamelCase = ViTForImageClassification(a__ ).eval() model.load_state_dict(a__ ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCamelCase = ViTImageProcessor() _UpperCamelCase = image_processor(images=prepare_img() , return_tensors="pt" ) _UpperCamelCase = encoding["pixel_values"] _UpperCamelCase = model(a__ ) if base_model: _UpperCamelCase = original_model(a__ ) assert torch.allclose(a__ , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: _UpperCamelCase = original_model(a__ ) assert logits.shape == outputs.logits.shape assert torch.allclose(a__ , outputs.logits , atol=1e-3 ) Path(a__ ).mkdir(exist_ok=a__ ) print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(a__ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) lowerCamelCase__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
701
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = MODEL_FOR_MASKED_LM_MAPPING __A = TF_MODEL_FOR_MASKED_LM_MAPPING def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-0_5, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Patrick", "score": 2.1e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-0_5, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCamelCase = unmasker("My name is <mask> <mask>" , top_k=2) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt") # convert model to fp16 pipe.model.half() _UpperCamelCase = pipe("Paris is the [MASK] of France.") # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_) @slow @require_torch def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt") self.run_large_test(lowercase_) @slow @require_tf def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf") self.run_large_test(lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> Any: """simple docstring""" _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is John", "score": 0.0_08, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_07, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_51, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_14, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is Patrick", "score": 0.0_05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_00, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_00, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int]) -> int: """simple docstring""" if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)") _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = fill_masker.tokenizer _UpperCamelCase = fill_masker.model _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , ) with self.assertRaises(lowercase_): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase_): fill_masker("This is") self.run_test_top_k(lowercase_ , lowercase_) self.run_test_targets(lowercase_ , lowercase_) self.run_test_top_k_targets(lowercase_ , lowercase_) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Dict , lowercase_ : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = sorted(vocab.keys())[:2] # Pipeline argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Call argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Score equivalence _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCamelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_) == set(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) # Raises with invalid with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""]) with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets="") def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : str , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) # top_k=2, ntargets=3 _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase_) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCamelCase = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_).issubset(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase_) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCamelCase = fill_masker(f'My name is {tokenizer.mask_token}' , targets=lowercase_ , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_) , 3) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , )
82
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( a__ ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = "huggingface/label-files" _UpperCamelCase = "imagenet-1k-id2label.json" _UpperCamelCase = json.load(open(hf_hub_download(a__ , a__ , repo_type="dataset" ) , "r" ) ) _UpperCamelCase = {int(a__ ): v for k, v in idalabel.items()} _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" _UpperCamelCase = BitConfig( conv_layer=a__ , num_labels=1_000 , idalabel=a__ , labelaid=a__ , ) return config def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' if "stem.conv" in name: _UpperCamelCase = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: _UpperCamelCase = name.replace("blocks" , "layers" ) if "head.fc" in name: _UpperCamelCase = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): _UpperCamelCase = "bit." + name if "bit" not in name and "classifier" not in name: _UpperCamelCase = "bit.encoder." + name return name def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCamelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( a__ , a__ , a__=False ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = get_config(a__ ) # load original model from timm _UpperCamelCase = create_model(a__ , pretrained=a__ ) timm_model.eval() # load state_dict of original model _UpperCamelCase = timm_model.state_dict() for key in state_dict.copy().keys(): _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val.squeeze() if "head" in key else val # load HuggingFace model _UpperCamelCase = BitForImageClassification(a__ ) model.eval() model.load_state_dict(a__ ) # create image processor _UpperCamelCase = create_transform(**resolve_data_config({} , model=a__ ) ) _UpperCamelCase = transform.transforms _UpperCamelCase = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } _UpperCamelCase = BitImageProcessor( do_resize=a__ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=a__ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=a__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) _UpperCamelCase = prepare_img() _UpperCamelCase = transform(a__ ).unsqueeze(0 ) _UpperCamelCase = processor(a__ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(a__ , a__ ) # verify logits with torch.no_grad(): _UpperCamelCase = model(a__ ) _UpperCamelCase = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) _UpperCamelCase = timm_model(a__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(a__ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(a__ ).mkdir(exist_ok=a__ ) print(f'Saving model {model_name} and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(a__ ) processor.save_pretrained(a__ ) if push_to_hub: print(f'Pushing model {model_name} and processor to the hub' ) model.push_to_hub(f'ybelkada/{model_name}' ) processor.push_to_hub(f'ybelkada/{model_name}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''resnetv2_50x1_bitm''', type=str, help='''Name of the BiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model to the hub.''', ) lowerCamelCase__ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
702
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json''', } class UpperCamelCase__ ( lowerCAmelCase ): '''simple docstring''' __A = '''mvp''' __A = ['''past_key_values'''] __A = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , lowercase_ : List[str]=50267 , lowercase_ : Any=1024 , lowercase_ : Dict=12 , lowercase_ : Any=4096 , lowercase_ : int=16 , lowercase_ : str=12 , lowercase_ : List[Any]=4096 , lowercase_ : List[str]=16 , lowercase_ : Dict=0.0 , lowercase_ : Any=0.0 , lowercase_ : Any="gelu" , lowercase_ : Optional[int]=1024 , lowercase_ : Tuple=0.1 , lowercase_ : List[str]=0.0 , lowercase_ : List[str]=0.0 , lowercase_ : Any=0.02 , lowercase_ : int=0.0 , lowercase_ : Optional[int]=False , lowercase_ : Any=True , lowercase_ : Optional[int]=1 , lowercase_ : Optional[int]=0 , lowercase_ : Any=2 , lowercase_ : str=True , lowercase_ : Dict=2 , lowercase_ : Dict=2 , lowercase_ : Union[str, Any]=False , lowercase_ : Optional[int]=100 , lowercase_ : List[Any]=800 , **lowercase_ : Union[str, Any] , ) -> List[str]: """simple docstring""" _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = d_model _UpperCamelCase = encoder_ffn_dim _UpperCamelCase = encoder_layers _UpperCamelCase = encoder_attention_heads _UpperCamelCase = decoder_ffn_dim _UpperCamelCase = decoder_layers _UpperCamelCase = decoder_attention_heads _UpperCamelCase = dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = activation_function _UpperCamelCase = init_std _UpperCamelCase = encoder_layerdrop _UpperCamelCase = decoder_layerdrop _UpperCamelCase = classifier_dropout _UpperCamelCase = use_cache _UpperCamelCase = encoder_layers _UpperCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase = use_prompt _UpperCamelCase = prompt_length _UpperCamelCase = prompt_mid_dim super().__init__( pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , decoder_start_token_id=lowercase_ , forced_eos_token_id=lowercase_ , **lowercase_ , ) if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , lowercase_): _UpperCamelCase = self.bos_token_id warnings.warn( f'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' "The config can simply be saved and uploaded again to be fixed.")
703
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
82
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_blip_2''': [ '''BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Blip2Config''', '''Blip2QFormerConfig''', '''Blip2VisionConfig''', ], '''processing_blip_2''': ['''Blip2Processor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Blip2Model''', '''Blip2QFormerModel''', '''Blip2PreTrainedModel''', '''Blip2ForConditionalGeneration''', '''Blip2VisionModel''', ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
704
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
0
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self : Dict) -> Optional[Any]: """simple docstring""" _UpperCamelCase = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") _UpperCamelCase = AutoTokenizer.from_pretrained("google/mt5-small") _UpperCamelCase = tokenizer("Hello there" , return_tensors="np").input_ids _UpperCamelCase = tokenizer("Hi I am" , return_tensors="np").input_ids _UpperCamelCase = shift_tokens_right(lowercase_ , model.config.pad_token_id , model.config.decoder_start_token_id) _UpperCamelCase = model(lowercase_ , decoder_input_ids=lowercase_).logits _UpperCamelCase = optax.softmax_cross_entropy(lowercase_ , onehot(lowercase_ , logits.shape[-1])).mean() _UpperCamelCase = -(labels.shape[-1] * loss.item()) _UpperCamelCase = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
705
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
82
0
import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = inspect.getfile(accelerate.test_utils) _UpperCamelCase = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ["scripts", "test_script.py"]) _UpperCamelCase = os.path.sep.join(inspect.getfile(self.__class__).split(os.path.sep)[:-1]) @require_tpu def __UpperCAmelCase ( self : List[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = f'\n {self.test_dir}/xla_spawn.py\n --num_cores 8\n {self.test_file_path}\n '.split() _UpperCamelCase = [sys.executable] + distributed_args execute_subprocess_async(lowercase_ , env=os.environ.copy())
706
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
82
0
def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: _UpperCamelCase = mf_knapsack(i - 1 , a__ , a__ , a__ ) else: _UpperCamelCase = max( mf_knapsack(i - 1 , a__ , a__ , a__ ) , mf_knapsack(i - 1 , a__ , a__ , j - wt[i - 1] ) + val[i - 1] , ) _UpperCamelCase = val return f[i][j] def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: _UpperCamelCase = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: _UpperCamelCase = dp[i - 1][w_] return dp[n][w_], dp def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Dict: '''simple docstring''' if not (isinstance(a__ , (list, tuple) ) and isinstance(a__ , (list, tuple) )): raise ValueError( "Both the weights and values vectors must be either lists or tuples" ) _UpperCamelCase = len(a__ ) if num_items != len(a__ ): _UpperCamelCase = ( "The number of weights must be the same as the number of values.\n" f'But got {num_items} weights and {len(a__ )} values' ) raise ValueError(a__ ) for i in range(a__ ): if not isinstance(wt[i] , a__ ): _UpperCamelCase = ( "All weights must be integers but got weight of " f'type {type(wt[i] )} at index {i}' ) raise TypeError(a__ ) _UpperCamelCase , _UpperCamelCase = knapsack(a__ , a__ , a__ , a__ ) _UpperCamelCase = set() _construct_solution(a__ , a__ , a__ , a__ , a__ ) return optimal_val, example_optional_set def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(a__ , a__ , i - 1 , a__ , a__ ) else: optimal_set.add(a__ ) _construct_solution(a__ , a__ , i - 1 , j - wt[i - 1] , a__ ) if __name__ == "__main__": lowerCamelCase__ = [3, 2, 4, 4] lowerCamelCase__ = [4, 3, 2, 3] lowerCamelCase__ = 4 lowerCamelCase__ = 6 lowerCamelCase__ = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] lowerCamelCase__,lowerCamelCase__ = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 lowerCamelCase__,lowerCamelCase__ = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print('''optimal_value = ''', optimal_solution) print('''An optimal subset corresponding to the optimal value''', optimal_subset)
707
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
82
0
from sklearn.metrics import fa_score import datasets lowerCamelCase__ = ''' The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) ''' lowerCamelCase__ = ''' Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`. - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives. - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {\'f1\': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results[\'f1\'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results[\'f1\'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro") >>> print(round(results[\'f1\'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {\'f1\': array([0.8, 0. , 0. ])} ''' lowerCamelCase__ = ''' @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("int32")), "references": datasets.Sequence(datasets.Value("int32")), } if self.config_name == "multilabel" else { "predictions": datasets.Value("int32"), "references": datasets.Value("int32"), }) , reference_urls=["https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"] , ) def __UpperCAmelCase ( self : List[str] , lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=None , lowercase_ : int=1 , lowercase_ : str="binary" , lowercase_ : List[Any]=None) -> List[Any]: """simple docstring""" _UpperCamelCase = fa_score( lowercase_ , lowercase_ , labels=lowercase_ , pos_label=lowercase_ , average=lowercase_ , sample_weight=lowercase_) return {"f1": float(lowercase_) if score.size == 1 else score}
708
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
82
0
import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''MCTCTFeatureExtractor''' __A = '''AutoTokenizer''' def __init__( self : Any , lowercase_ : Any , lowercase_ : List[str]) -> Dict: """simple docstring""" super().__init__(lowercase_ , lowercase_) _UpperCamelCase = self.feature_extractor _UpperCamelCase = False def __call__( self : Union[str, Any] , *lowercase_ : Union[str, Any] , **lowercase_ : List[str]) -> Any: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*lowercase_ , **lowercase_) if "raw_speech" in kwargs: warnings.warn("Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.") _UpperCamelCase = kwargs.pop("raw_speech") else: _UpperCamelCase = kwargs.pop("audio" , lowercase_) _UpperCamelCase = kwargs.pop("sampling_rate" , lowercase_) _UpperCamelCase = kwargs.pop("text" , lowercase_) if len(lowercase_) > 0: _UpperCamelCase = args[0] _UpperCamelCase = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process.") if audio is not None: _UpperCamelCase = self.feature_extractor(lowercase_ , *lowercase_ , sampling_rate=lowercase_ , **lowercase_) if text is not None: _UpperCamelCase = self.tokenizer(lowercase_ , **lowercase_) if text is None: return inputs elif audio is None: return encodings else: _UpperCamelCase = encodings["input_ids"] return inputs def __UpperCAmelCase ( self : Optional[int] , *lowercase_ : int , **lowercase_ : List[Any]) -> Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , *lowercase_ : List[Any] , **lowercase_ : List[Any]) -> List[str]: """simple docstring""" if self._in_target_context_manager: return self.current_processor.pad(*lowercase_ , **lowercase_) _UpperCamelCase = kwargs.pop("input_features" , lowercase_) _UpperCamelCase = kwargs.pop("labels" , lowercase_) if len(lowercase_) > 0: _UpperCamelCase = args[0] _UpperCamelCase = args[1:] if input_features is not None: _UpperCamelCase = self.feature_extractor.pad(lowercase_ , *lowercase_ , **lowercase_) if labels is not None: _UpperCamelCase = self.tokenizer.pad(lowercase_ , **lowercase_) if labels is None: return input_features elif input_features is None: return labels else: _UpperCamelCase = labels["input_ids"] return input_features def __UpperCAmelCase ( self : List[str] , *lowercase_ : str , **lowercase_ : Optional[Any]) -> Optional[int]: """simple docstring""" return self.tokenizer.decode(*lowercase_ , **lowercase_) @contextmanager def __UpperCAmelCase ( self : Dict) -> Tuple: """simple docstring""" warnings.warn( "`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your " "labels by using the argument `text` of the regular `__call__` method (either in the same call as " "your audio inputs, or in a separate call.") _UpperCamelCase = True _UpperCamelCase = self.tokenizer yield _UpperCamelCase = self.feature_extractor _UpperCamelCase = False
709
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , *lowercase_ : List[str] , **lowercase_ : Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
82
0
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class _UpperCAmelCase ( unittest.TestCase, lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[Any]) -> Any: """simple docstring""" _UpperCamelCase = load_tool("text-classification") self.tool.setup() _UpperCamelCase = load_tool("text-classification" , remote=lowercase_) def __UpperCAmelCase ( self : Optional[Any]) -> Any: """simple docstring""" _UpperCamelCase = self.tool("That's quite cool" , ["positive", "negative"]) self.assertEqual(lowercase_ , "positive") def __UpperCAmelCase ( self : Optional[int]) -> Dict: """simple docstring""" _UpperCamelCase = self.remote_tool("That's quite cool" , ["positive", "negative"]) self.assertEqual(lowercase_ , "positive") def __UpperCAmelCase ( self : List[str]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.tool(text="That's quite cool" , labels=["positive", "negative"]) self.assertEqual(lowercase_ , "positive") def __UpperCAmelCase ( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.remote_tool(text="That's quite cool" , labels=["positive", "negative"]) self.assertEqual(lowercase_ , "positive")
710
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
82
0
'''simple docstring''' from __future__ import annotations from collections import deque class _UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : list[str]) -> Optional[int]: """simple docstring""" _UpperCamelCase = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []}) for keyword in keywords: self.add_keyword(lowercase_) self.set_fail_transitions() def __UpperCAmelCase ( self : Tuple , lowercase_ : int , lowercase_ : str) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def __UpperCAmelCase ( self : Optional[int] , lowercase_ : str) -> None: """simple docstring""" _UpperCamelCase = 0 for character in keyword: _UpperCamelCase = self.find_next_state(lowercase_ , lowercase_) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], }) self.adlist[current_state]["next_states"].append(len(self.adlist) - 1) _UpperCamelCase = len(self.adlist) - 1 else: _UpperCamelCase = next_state self.adlist[current_state]["output"].append(lowercase_) def __UpperCAmelCase ( self : Tuple) -> None: """simple docstring""" _UpperCamelCase = deque() for node in self.adlist[0]["next_states"]: q.append(lowercase_) _UpperCamelCase = 0 while q: _UpperCamelCase = q.popleft() for child in self.adlist[r]["next_states"]: q.append(lowercase_) _UpperCamelCase = self.adlist[r]["fail_state"] while ( self.find_next_state(lowercase_ , self.adlist[child]["value"]) is None and state != 0 ): _UpperCamelCase = self.adlist[state]["fail_state"] _UpperCamelCase = self.find_next_state( lowercase_ , self.adlist[child]["value"]) if self.adlist[child]["fail_state"] is None: _UpperCamelCase = 0 _UpperCamelCase = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def __UpperCAmelCase ( self : int , lowercase_ : str) -> dict[str, list[int]]: """simple docstring""" _UpperCamelCase = {} # returns a dict with keywords and list of its occurrences _UpperCamelCase = 0 for i in range(len(lowercase_)): while ( self.find_next_state(lowercase_ , string[i]) is None and current_state != 0 ): _UpperCamelCase = self.adlist[current_state]["fail_state"] _UpperCamelCase = self.find_next_state(lowercase_ , string[i]) if next_state is None: _UpperCamelCase = 0 else: _UpperCamelCase = next_state for key in self.adlist[current_state]["output"]: if key not in result: _UpperCamelCase = [] result[key].append(i - len(lowercase_) + 1) return result if __name__ == "__main__": import doctest doctest.testmod()
711
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
82
0
import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Any , lowercase_ : str) -> Dict: """simple docstring""" _UpperCamelCase = parent def __UpperCAmelCase ( self : Union[str, Any]) -> Dict: """simple docstring""" return {} def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = "<HTML>\n\n <HEAD>\n <TITLE>sample document</TITLE>\n </HEAD>\n\n <BODY BGCOLOR=\"FFFFFF\">\n <HR>\n <a href=\"http://google.com\">Goog</a>\n <H1>This is one header</H1>\n <H2>This is a another Header</H2>\n <P>Travel from\n <P>\n <B>SFO to JFK</B>\n <BR>\n <B><I>on May 2, 2015 at 2:00 pm. For details go to confirm.com </I></B>\n <HR>\n <div style=\"color:#0000FF\">\n <h3>Traveler <b> name </b> is\n <p> John Doe </p>\n </div>" _UpperCamelCase = "\n <!DOCTYPE html>\n <html>\n <body>\n\n <h1>My First Heading</h1>\n <p>My first paragraph.</p>\n\n </body>\n </html>\n " return [html_string_a, html_string_a] @require_bsa class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = MarkupLMFeatureExtractor if is_bsa_available() else None def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = MarkupLMFeatureExtractionTester(self) @property def __UpperCAmelCase ( self : Optional[Any]) -> str: """simple docstring""" return self.feature_extract_tester.prepare_feat_extract_dict() def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.feature_extraction_class() # Test not batched input _UpperCamelCase = get_html_strings()[0] _UpperCamelCase = feature_extractor(lowercase_) # fmt: off _UpperCamelCase = [["sample document", "Goog", "This is one header", "This is a another Header", "Travel from", "SFO to JFK", "on May 2, 2015 at 2:00 pm. For details go to confirm.com", "Traveler", "name", "is", "John Doe"]] _UpperCamelCase = [["/html/head/title", "/html/body/a", "/html/body/h1", "/html/body/h2", "/html/body/p", "/html/body/p/p/b[1]", "/html/body/p/p/b[2]/i", "/html/body/p/p/div/h3", "/html/body/p/p/div/h3/b", "/html/body/p/p/div/h3", "/html/body/p/p/div/h3/p"]] # fmt: on self.assertEqual(encoding.nodes , lowercase_) self.assertEqual(encoding.xpaths , lowercase_) # Test batched _UpperCamelCase = get_html_strings() _UpperCamelCase = feature_extractor(lowercase_) # fmt: off _UpperCamelCase = expected_nodes + [["My First Heading", "My first paragraph."]] _UpperCamelCase = expected_xpaths + [["/html/body/h1", "/html/body/p"]] self.assertEqual(len(encoding.nodes) , 2) self.assertEqual(len(encoding.xpaths) , 2) self.assertEqual(encoding.nodes , lowercase_) self.assertEqual(encoding.xpaths , lowercase_)
712
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
82
0
def lowerCAmelCase__ ( a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = "" for word_or_phrase in separated: if not isinstance(a__ , a__ ): raise Exception("join() accepts only strings to be joined" ) joined += word_or_phrase + separator return joined.strip(a__ ) if __name__ == "__main__": from doctest import testmod testmod()
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCamelCase__ = {'''configuration_deit''': ['''DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''DeiTConfig''', '''DeiTOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['''DeiTFeatureExtractor'''] lowerCamelCase__ = ['''DeiTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''DEIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''DeiTForImageClassification''', '''DeiTForImageClassificationWithTeacher''', '''DeiTForMaskedImageModeling''', '''DeiTModel''', '''DeiTPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFDeiTForImageClassification''', '''TFDeiTForImageClassificationWithTeacher''', '''TFDeiTForMaskedImageModeling''', '''TFDeiTModel''', '''TFDeiTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
def lowerCAmelCase__ ( a__ ) ->list[int]: '''simple docstring''' if length <= 0 or not isinstance(a__ , a__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(a__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
import datasets from .evaluate import evaluate lowerCamelCase__ = '''\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } ''' lowerCamelCase__ = ''' This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. ''' lowerCamelCase__ = ''' Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': the text of the answer references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the SQuAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}] >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}] >>> squad_metric = datasets.load_metric("squad") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": {"id": datasets.Value("string"), "prediction_text": datasets.Value("string")}, "references": { "id": datasets.Value("string"), "answers": datasets.features.Sequence( { "text": datasets.Value("string"), "answer_start": datasets.Value("int32"), }), }, }) , codebase_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , reference_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : int , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = {prediction["id"]: prediction["prediction_text"] for prediction in predictions} _UpperCamelCase = [ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] _UpperCamelCase = evaluate(dataset=lowercase_ , predictions=lowercase_) return score
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): lowerCamelCase__ = True from torch.cuda.amp import autocast lowerCamelCase__ = logging.getLogger(__name__) @dataclass class _UpperCAmelCase : '''simple docstring''' __A = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''}, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Whether to log verbose messages or not.'''}, ) __A = field( default=2.0, metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) __A = field( default=0.5, metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) __A = field( default=0.999_995, metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def lowerCAmelCase__ ( a__ , a__ ) ->Any: '''simple docstring''' logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) _UpperCamelCase = logging.WARNING if model_args.verbose_logging: _UpperCamelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): _UpperCamelCase = logging.INFO logger.setLevel(a__ ) @dataclass class _UpperCAmelCase : '''simple docstring''' __A = field( default=lowerCAmelCase, metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) __A = field( default='''train''', metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' }, ) __A = field( default='''validation''', metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) }, ) __A = field( default='''file''', metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''}, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) __A = field( default=1, metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' }, ) __A = field( default=lowerCAmelCase, metadata={'''help''': '''The number of processes to use for the preprocessing.'''}, ) __A = field( default=20.0, metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class _UpperCAmelCase : '''simple docstring''' __A = 42 __A = 42 __A = '''longest''' __A = None __A = None def __call__( self : Tuple , lowercase_ : List[Dict[str, Union[List[int], torch.Tensor]]]) -> Dict[str, torch.Tensor]: """simple docstring""" _UpperCamelCase = self.feature_extractor.pad( lowercase_ , max_length=self.max_length , padding=self.padding , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" , ) _UpperCamelCase = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1]) _UpperCamelCase = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula _UpperCamelCase = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1)).to( torch.long) _UpperCamelCase = torch.zeros( (batch_size, mask_indices_seq_length) , dtype=torch.long , device=batch["input_values"].device) # these two operations makes sure that all values # before the output lengths indices are attended to _UpperCamelCase = 1 _UpperCamelCase = attention_mask.flip([-1]).cumsum(-1).flip([-1]).bool() # sample randomly masked indices _UpperCamelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) , self.model.config.mask_time_prob , self.model.config.mask_time_length , attention_mask=lowercase_ , min_masks=2 , ) return batch class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Tuple , *lowercase_ : Dict , lowercase_ : Optional[int]=1 , lowercase_ : Union[str, Any]=0 , lowercase_ : Tuple=1.0 , **lowercase_ : Tuple) -> Tuple: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) _UpperCamelCase = 0 _UpperCamelCase = max_gumbel_temp _UpperCamelCase = min_gumbel_temp _UpperCamelCase = gumbel_temp_decay def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : nn.Module , lowercase_ : Dict[str, Union[torch.Tensor, Any]]) -> torch.Tensor: """simple docstring""" model.train() _UpperCamelCase = self._prepare_inputs(lowercase_) if self.use_amp: with autocast(): _UpperCamelCase = self.compute_loss(lowercase_ , lowercase_) else: _UpperCamelCase = self.compute_loss(lowercase_ , lowercase_) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": _UpperCamelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": _UpperCamelCase = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']') if self.args.gradient_accumulation_steps > 1: _UpperCamelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowercase_).backward() elif self.use_apex: with amp.scale_loss(lowercase_ , self.optimizer) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowercase_) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp)) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp)) return loss.detach() def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = parser.parse_args_into_dataclasses() configure_logger(a__ , a__ ) # Downloading and loading a dataset from the hub. _UpperCamelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" _UpperCamelCase = DatasetDict() _UpperCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'{data_args.train_split_name}[:{data_args.validation_split_percentage}%]' , cache_dir=model_args.cache_dir , ) _UpperCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'{data_args.train_split_name}[{data_args.validation_split_percentage}%:]' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" _UpperCamelCase = DatasetDict() _UpperCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) _UpperCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'{data_args.train_split_name}' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported _UpperCamelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=a__ ) def prepare_dataset(a__ ): # check that all files have the correct sampling rate _UpperCamelCase , _UpperCamelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays _UpperCamelCase = datasets.map( a__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names ) # filter audio files that are too long _UpperCamelCase = vectorized_datasets.filter( lambda a__ : len(data["speech"] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(a__ ): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` _UpperCamelCase = vectorized_datasets.map( a__ , batched=a__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 _UpperCamelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'" ) _UpperCamelCase = WavaVecaForPreTraining(a__ ) _UpperCamelCase = DataCollatorForWavaVecaPretraining(model=a__ , feature_extractor=a__ ) _UpperCamelCase = WavaVecaPreTrainer( model=a__ , data_collator=a__ , args=a__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=a__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( a__ ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=["stage2", "stage3", "stage4"] , ) _UpperCamelCase = DetaConfig( backbone_config=a__ , num_queries=900 , encoder_ffn_dim=2_048 , decoder_ffn_dim=2_048 , num_feature_levels=5 , assign_first_stage=a__ , with_box_refine=a__ , two_stage=a__ , ) # set labels _UpperCamelCase = "huggingface/label-files" if "o365" in model_name: _UpperCamelCase = 366 _UpperCamelCase = "object365-id2label.json" else: _UpperCamelCase = 91 _UpperCamelCase = "coco-detection-id2label.json" _UpperCamelCase = num_labels _UpperCamelCase = json.load(open(cached_download(hf_hub_url(a__ , a__ , repo_type="dataset" ) ) , "r" ) ) _UpperCamelCase = {int(a__ ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = [] # stem # fmt: off rename_keys.append(("backbone.0.body.patch_embed.proj.weight", "model.backbone.model.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("backbone.0.body.patch_embed.proj.bias", "model.backbone.model.embeddings.patch_embeddings.projection.bias") ) rename_keys.append(("backbone.0.body.patch_embed.norm.weight", "model.backbone.model.embeddings.norm.weight") ) rename_keys.append(("backbone.0.body.patch_embed.norm.bias", "model.backbone.model.embeddings.norm.bias") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias') ) if i < 3: rename_keys.append((f'backbone.0.body.layers.{i}.downsample.reduction.weight', f'model.backbone.model.encoder.layers.{i}.downsample.reduction.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.weight', f'model.backbone.model.encoder.layers.{i}.downsample.norm.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.bias', f'model.backbone.model.encoder.layers.{i}.downsample.norm.bias') ) rename_keys.append(("backbone.0.body.norm1.weight", "model.backbone.model.hidden_states_norms.stage2.weight") ) rename_keys.append(("backbone.0.body.norm1.bias", "model.backbone.model.hidden_states_norms.stage2.bias") ) rename_keys.append(("backbone.0.body.norm2.weight", "model.backbone.model.hidden_states_norms.stage3.weight") ) rename_keys.append(("backbone.0.body.norm2.bias", "model.backbone.model.hidden_states_norms.stage3.bias") ) rename_keys.append(("backbone.0.body.norm3.weight", "model.backbone.model.hidden_states_norms.stage4.weight") ) rename_keys.append(("backbone.0.body.norm3.bias", "model.backbone.model.hidden_states_norms.stage4.bias") ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight', f'model.encoder.layers.{i}.self_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias', f'model.encoder.layers.{i}.self_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.weight', f'model.encoder.layers.{i}.self_attn.attention_weights.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.bias', f'model.encoder.layers.{i}.self_attn.attention_weights.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.weight', f'model.encoder.layers.{i}.self_attn.value_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.bias', f'model.encoder.layers.{i}.self_attn.value_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.weight', f'model.encoder.layers.{i}.self_attn.output_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.bias', f'model.encoder.layers.{i}.self_attn.output_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.weight', f'model.encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'model.encoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'model.encoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'model.encoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'model.encoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'model.encoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'model.encoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'model.encoder.layers.{i}.final_layer_norm.bias') ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.weight', f'model.decoder.layers.{i}.encoder_attn.attention_weights.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.bias', f'model.decoder.layers.{i}.encoder_attn.attention_weights.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.weight', f'model.decoder.layers.{i}.encoder_attn.value_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.bias', f'model.decoder.layers.{i}.encoder_attn.value_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.weight', f'model.decoder.layers.{i}.encoder_attn.output_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.bias', f'model.decoder.layers.{i}.encoder_attn.output_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.weight', f'model.decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'model.decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'model.decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'model.decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.weight', f'model.decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.bias', f'model.decoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'model.decoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'model.decoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'model.decoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'model.decoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'model.decoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'model.decoder.layers.{i}.final_layer_norm.bias') ) # fmt: on return rename_keys def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = dct.pop(a__ ) _UpperCamelCase = val def lowerCAmelCase__ ( a__ , a__ ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): _UpperCamelCase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight' ) _UpperCamelCase = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[:dim, :] _UpperCamelCase = in_proj_bias[: dim] _UpperCamelCase = in_proj_weight[ dim : dim * 2, : ] _UpperCamelCase = in_proj_bias[ dim : dim * 2 ] _UpperCamelCase = in_proj_weight[ -dim :, : ] _UpperCamelCase = in_proj_bias[-dim :] # fmt: on def lowerCAmelCase__ ( a__ , a__ ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention _UpperCamelCase = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _UpperCamelCase = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[:hidden_size, :] _UpperCamelCase = in_proj_bias[:hidden_size] _UpperCamelCase = in_proj_weight[ hidden_size : hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[hidden_size : hidden_size * 2] _UpperCamelCase = in_proj_weight[-hidden_size:, :] _UpperCamelCase = in_proj_bias[-hidden_size:] def lowerCAmelCase__ ( ) ->str: '''simple docstring''' _UpperCamelCase = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCamelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Tuple: '''simple docstring''' _UpperCamelCase = get_deta_config(a__ ) # load original state dict if model_name == "deta-swin-large": _UpperCamelCase = hf_hub_download(repo_id="nielsr/deta-checkpoints" , filename="adet_swin_ft.pth" ) elif model_name == "deta-swin-large-o365": _UpperCamelCase = hf_hub_download(repo_id="jozhang97/deta-swin-l-o365" , filename="deta_swin_pt_o365.pth" ) else: raise ValueError(f'Model name {model_name} not supported' ) _UpperCamelCase = torch.load(a__ , map_location="cpu" )["model"] # original state dict for name, param in state_dict.items(): print(a__ , param.shape ) # rename keys _UpperCamelCase = create_rename_keys(a__ ) for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) read_in_swin_q_k_v(a__ , config.backbone_config ) read_in_decoder_q_k_v(a__ , a__ ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val if "input_proj" in key: _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: _UpperCamelCase = state_dict.pop(a__ ) _UpperCamelCase = val # finally, create HuggingFace model and load state dict _UpperCamelCase = DetaForObjectDetection(a__ ) model.load_state_dict(a__ ) model.eval() _UpperCamelCase = "cuda" if torch.cuda.is_available() else "cpu" model.to(a__ ) # load image processor _UpperCamelCase = DetaImageProcessor(format="coco_detection" ) # verify our conversion on image _UpperCamelCase = prepare_img() _UpperCamelCase = processor(images=a__ , return_tensors="pt" ) _UpperCamelCase = encoding["pixel_values"] _UpperCamelCase = model(pixel_values.to(a__ ) ) # verify logits print("Logits:" , outputs.logits[0, :3, :3] ) print("Boxes:" , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": _UpperCamelCase = torch.tensor( [[-7.6308, -2.8485, -5.3737], [-7.2037, -4.5505, -4.8027], [-7.2943, -4.2611, -4.6617]] ) _UpperCamelCase = torch.tensor([[0.4987, 0.4969, 0.9999], [0.2549, 0.5498, 0.4805], [0.5498, 0.2757, 0.0569]] ) elif model_name == "deta-swin-large-o365": _UpperCamelCase = torch.tensor( [[-8.0122, -3.5720, -4.9717], [-8.1547, -3.6886, -4.6389], [-7.6610, -3.6194, -5.0134]] ) _UpperCamelCase = torch.tensor([[0.2523, 0.5549, 0.4881], [0.7715, 0.4149, 0.4601], [0.5503, 0.2753, 0.0575]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(a__ ) , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(a__ ) , atol=1e-4 ) print("Everything ok!" ) if pytorch_dump_folder_path: # Save model and processor logger.info(f'Saving PyTorch model and processor to {pytorch_dump_folder_path}...' ) Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) processor.save_pretrained(a__ ) # Push to hub if push_to_hub: print("Pushing model and processor to hub..." ) model.push_to_hub(f'jozhang97/{model_name}' ) processor.push_to_hub(f'jozhang97/{model_name}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCamelCase__ = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
700
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
82
0
from collections.abc import Sequence def lowerCAmelCase__ ( a__ , a__ ) ->float: '''simple docstring''' return sum(c * (x**i) for i, c in enumerate(a__ ) ) def lowerCAmelCase__ ( a__ , a__ ) ->float: '''simple docstring''' _UpperCamelCase = 0.0 for coeff in reversed(a__ ): _UpperCamelCase = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
701
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = MODEL_FOR_MASKED_LM_MAPPING __A = TF_MODEL_FOR_MASKED_LM_MAPPING def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-0_5, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Patrick", "score": 2.1e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-0_5, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCamelCase = unmasker("My name is <mask> <mask>" , top_k=2) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt") # convert model to fp16 pipe.model.half() _UpperCamelCase = pipe("Paris is the [MASK] of France.") # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_) @slow @require_torch def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt") self.run_large_test(lowercase_) @slow @require_tf def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf") self.run_large_test(lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> Any: """simple docstring""" _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is John", "score": 0.0_08, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_07, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_51, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_14, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is Patrick", "score": 0.0_05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_00, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_00, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int]) -> int: """simple docstring""" if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)") _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = fill_masker.tokenizer _UpperCamelCase = fill_masker.model _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , ) with self.assertRaises(lowercase_): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase_): fill_masker("This is") self.run_test_top_k(lowercase_ , lowercase_) self.run_test_targets(lowercase_ , lowercase_) self.run_test_top_k_targets(lowercase_ , lowercase_) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Dict , lowercase_ : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = sorted(vocab.keys())[:2] # Pipeline argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Call argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Score equivalence _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCamelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_) == set(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) # Raises with invalid with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""]) with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets="") def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : str , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) # top_k=2, ntargets=3 _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase_) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCamelCase = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_).issubset(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase_) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCamelCase = fill_masker(f'My name is {tokenizer.mask_token}' , targets=lowercase_ , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_) , 3) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , )
82
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
702
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class UpperCamelCase__ ( lowerCAmelCase ): '''simple docstring''' __A = '''vit_msn''' def __init__( self : Union[str, Any] , lowercase_ : Optional[int]=768 , lowercase_ : int=12 , lowercase_ : Any=12 , lowercase_ : int=3072 , lowercase_ : Optional[Any]="gelu" , lowercase_ : str=0.0 , lowercase_ : Any=0.0 , lowercase_ : Tuple=0.02 , lowercase_ : Optional[Any]=1e-0_6 , lowercase_ : Union[str, Any]=224 , lowercase_ : Optional[int]=16 , lowercase_ : List[str]=3 , lowercase_ : List[Any]=True , **lowercase_ : str , ) -> str: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = qkv_bias
703
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
82
0
from __future__ import annotations def lowerCAmelCase__ ( a__ ) ->float: '''simple docstring''' if not nums: raise ValueError("List is empty" ) return sum(a__ ) / len(a__ ) if __name__ == "__main__": import doctest doctest.testmod()
704
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
0
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
705
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
82
0
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class SCREAMING_SNAKE_CASE ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
706
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
82
0
# This is the module that test_patching.py uses to test patch_submodule() import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests lowerCamelCase__ = open # noqa: we just need to have a builtin inside this module to test it properly
707
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
82
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} # See all MVP models at https://huggingface.co/models?filter=mvp lowerCamelCase__ = { '''vocab_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json''', }, '''added_tokens.json''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json''', }, '''merges_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json''', }, } lowerCamelCase__ = { '''RUCAIBox/mvp''': 1024, } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] __A = MvpTokenizer def __init__( self : Union[str, Any] , lowercase_ : Optional[int]=None , lowercase_ : List[str]=None , lowercase_ : str=None , lowercase_ : Union[str, Any]="replace" , lowercase_ : List[Any]="<s>" , lowercase_ : Optional[int]="</s>" , lowercase_ : Tuple="</s>" , lowercase_ : Tuple="<s>" , lowercase_ : List[Any]="<unk>" , lowercase_ : Union[str, Any]="<pad>" , lowercase_ : Optional[int]="<mask>" , lowercase_ : List[str]=False , lowercase_ : List[str]=True , **lowercase_ : str , ) -> str: """simple docstring""" super().__init__( lowercase_ , lowercase_ , tokenizer_file=lowercase_ , errors=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , mask_token=lowercase_ , add_prefix_space=lowercase_ , trim_offsets=lowercase_ , **lowercase_ , ) _UpperCamelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , lowercase_) != add_prefix_space: _UpperCamelCase = getattr(lowercase_ , pre_tok_state.pop("type")) _UpperCamelCase = add_prefix_space _UpperCamelCase = pre_tok_class(**lowercase_) _UpperCamelCase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` _UpperCamelCase = "post_processor" _UpperCamelCase = getattr(self.backend_tokenizer , lowercase_ , lowercase_) if tokenizer_component_instance: _UpperCamelCase = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _UpperCamelCase = tuple(state["sep"]) if "cls" in state: _UpperCamelCase = tuple(state["cls"]) _UpperCamelCase = False if state.get("add_prefix_space" , lowercase_) != add_prefix_space: _UpperCamelCase = add_prefix_space _UpperCamelCase = True if state.get("trim_offsets" , lowercase_) != trim_offsets: _UpperCamelCase = trim_offsets _UpperCamelCase = True if changes_to_apply: _UpperCamelCase = getattr(lowercase_ , state.pop("type")) _UpperCamelCase = component_class(**lowercase_) setattr(self.backend_tokenizer , lowercase_ , lowercase_) @property def __UpperCAmelCase ( self : List[str]) -> str: """simple docstring""" if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet.") return None return str(self._mask_token) @mask_token.setter def __UpperCAmelCase ( self : Any , lowercase_ : Tuple) -> Any: """simple docstring""" _UpperCamelCase = AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_) if isinstance(lowercase_ , lowercase_) else value _UpperCamelCase = value def __UpperCAmelCase ( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : Union[str, Any]) -> BatchEncoding: """simple docstring""" _UpperCamelCase = kwargs.get("is_split_into_words" , lowercase_) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs.") return super()._batch_encode_plus(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Optional[int] , *lowercase_ : Optional[int] , **lowercase_ : Dict) -> BatchEncoding: """simple docstring""" _UpperCamelCase = kwargs.get("is_split_into_words" , lowercase_) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs.") return super()._encode_plus(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : str , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" _UpperCamelCase = self._tokenizer.model.save(lowercase_ , name=lowercase_) return tuple(lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Dict , lowercase_ : Union[str, Any]=None) -> Dict: """simple docstring""" _UpperCamelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCAmelCase ( self : str , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
708
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
82
0
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCamelCase__ = get_logger(__name__) lowerCamelCase__ = R''' Args: input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam search or log softmax for each vocabulary token when using beam search kwargs (`Dict[str, Any]`, *optional*): Additional logits processor specific kwargs. Return: `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores. ''' class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(lowercase_) def __call__( self : List[Any] , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(lowercase_) def __call__( self : Tuple , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.') class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' @add_start_docstrings(lowercase_) def __call__( self : Optional[int] , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int , **lowercase_ : Dict) -> jnp.ndarray: """simple docstring""" for processor in self: _UpperCamelCase = inspect.signature(processor.__call__).parameters if len(lowercase_) > 3: if not all(arg in kwargs for arg in list(function_args.keys())[2:]): raise ValueError( f'Make sure that all the required parameters: {list(function_args.keys())} for ' f'{processor.__class__} are passed to the logits processor.') _UpperCamelCase = processor(lowercase_ , lowercase_ , lowercase_ , **lowercase_) else: _UpperCamelCase = processor(lowercase_ , lowercase_ , lowercase_) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : float) -> Tuple: """simple docstring""" if not isinstance(lowercase_ , lowercase_) or not (temperature > 0): raise ValueError(f'`temperature` has to be a strictly positive float, but is {temperature}') _UpperCamelCase = temperature def __call__( self : Any , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase = scores / self.temperature return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[str] , lowercase_ : float , lowercase_ : float = -float("Inf") , lowercase_ : int = 1) -> Optional[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_) or (top_p < 0 or top_p > 1.0): raise ValueError(f'`top_p` has to be a float > 0 and < 1, but is {top_p}') if not isinstance(lowercase_ , lowercase_) or (min_tokens_to_keep < 1): raise ValueError(f'`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}') _UpperCamelCase = top_p _UpperCamelCase = filter_value _UpperCamelCase = min_tokens_to_keep def __call__( self : Any , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase , _UpperCamelCase = lax.top_k(lowercase_ , scores.shape[-1]) _UpperCamelCase = jnp.full_like(lowercase_ , self.filter_value) _UpperCamelCase = jax.nn.softmax(lowercase_ , axis=-1).cumsum(axis=-1) _UpperCamelCase = cumulative_probs < self.top_p # include the token that is higher than top_p as well _UpperCamelCase = jnp.roll(lowercase_ , 1) score_mask |= score_mask.at[:, 0].set(lowercase_) # min tokens to keep _UpperCamelCase = score_mask.at[:, : self.min_tokens_to_keep].set(lowercase_) _UpperCamelCase = jnp.where(lowercase_ , lowercase_ , lowercase_) _UpperCamelCase = jax.lax.sort_key_val(lowercase_ , lowercase_)[-1] return next_scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[str] , lowercase_ : int , lowercase_ : float = -float("Inf") , lowercase_ : int = 1) -> Any: """simple docstring""" if not isinstance(lowercase_ , lowercase_) or top_k <= 0: raise ValueError(f'`top_k` has to be a strictly positive integer, but is {top_k}') _UpperCamelCase = max(lowercase_ , lowercase_) _UpperCamelCase = filter_value def __call__( self : List[Any] , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase , _UpperCamelCase = scores.shape _UpperCamelCase = jnp.full(batch_size * vocab_size , self.filter_value) _UpperCamelCase = min(self.top_k , scores.shape[-1]) # Safety check _UpperCamelCase , _UpperCamelCase = lax.top_k(lowercase_ , lowercase_) _UpperCamelCase = jnp.broadcast_to((jnp.arange(lowercase_) * vocab_size)[:, None] , (batch_size, topk)).flatten() _UpperCamelCase = topk_scores.flatten() _UpperCamelCase = topk_indices.flatten() + shift _UpperCamelCase = next_scores_flat.at[topk_indices_flat].set(lowercase_) _UpperCamelCase = next_scores_flat.reshape(lowercase_ , lowercase_) return next_scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , lowercase_ : int) -> Optional[Any]: """simple docstring""" _UpperCamelCase = bos_token_id def __call__( self : Union[str, Any] , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase = jnp.full(scores.shape , -float("inf")) _UpperCamelCase = 1 - jnp.bool_(cur_len - 1) _UpperCamelCase = jnp.where(lowercase_ , new_scores.at[:, self.bos_token_id].set(0) , lowercase_) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase_ : int , lowercase_ : int) -> Tuple: """simple docstring""" _UpperCamelCase = max_length _UpperCamelCase = eos_token_id def __call__( self : str , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase = jnp.full(scores.shape , -float("inf")) _UpperCamelCase = 1 - jnp.bool_(cur_len - self.max_length + 1) _UpperCamelCase = jnp.where(lowercase_ , new_scores.at[:, self.eos_token_id].set(0) , lowercase_) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , lowercase_ : int , lowercase_ : int) -> Dict: """simple docstring""" if not isinstance(lowercase_ , lowercase_) or min_length < 0: raise ValueError(f'`min_length` has to be a positive integer, but is {min_length}') if not isinstance(lowercase_ , lowercase_) or eos_token_id < 0: raise ValueError(f'`eos_token_id` has to be a positive integer, but is {eos_token_id}') _UpperCamelCase = min_length _UpperCamelCase = eos_token_id def __call__( self : Tuple , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase = 1 - jnp.clip(cur_len - self.min_length , 0 , 1) _UpperCamelCase = jnp.where(lowercase_ , scores.at[:, self.eos_token_id].set(-float("inf")) , lowercase_) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Tuple , lowercase_ : List[Any] , lowercase_ : int) -> Tuple: """simple docstring""" _UpperCamelCase = list(lowercase_) _UpperCamelCase = begin_index def __call__( self : Tuple , lowercase_ : int , lowercase_ : Dict , lowercase_ : int) -> Optional[Any]: """simple docstring""" _UpperCamelCase = 1 - jnp.bool_(cur_len - self.begin_index) _UpperCamelCase = jnp.where(lowercase_ , scores.at[:, self.begin_suppress_tokens].set(-float("inf")) , lowercase_) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : list) -> Optional[Any]: """simple docstring""" _UpperCamelCase = list(lowercase_) def __call__( self : Union[str, Any] , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" _UpperCamelCase = scores.at[..., self.suppress_tokens].set(-float("inf")) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , lowercase_ : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = dict(lowercase_) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. _UpperCamelCase = jnp.ones((max(force_token_map.keys()) + 1) , dtype=jnp.intaa) * -1 for index, token in force_token_map.items(): if token is not None: _UpperCamelCase = force_token_array.at[index].set(lowercase_) _UpperCamelCase = jnp.intaa(lowercase_) def __call__( self : Tuple , lowercase_ : jnp.ndarray , lowercase_ : jnp.ndarray , lowercase_ : int) -> jnp.ndarray: """simple docstring""" def _force_token(lowercase_ : Optional[Any]): _UpperCamelCase = scores.shape[0] _UpperCamelCase = self.force_token_array[generation_idx] _UpperCamelCase = jnp.ones_like(lowercase_ , dtype=scores.dtype) * -float("inf") _UpperCamelCase = jnp.zeros((batch_size, 1) , dtype=scores.dtype) _UpperCamelCase = lax.dynamic_update_slice(lowercase_ , lowercase_ , (0, current_token)) return new_scores _UpperCamelCase = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(lowercase_) , lambda: scores , ) , ) return scores class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[int] , lowercase_ : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = generate_config.eos_token_id _UpperCamelCase = generate_config.no_timestamps_token_id _UpperCamelCase = generate_config.no_timestamps_token_id + 1 _UpperCamelCase = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(lowercase_ , "max_initial_timestamp_index"): _UpperCamelCase = generate_config.max_initial_timestamp_index else: _UpperCamelCase = model_config.vocab_size if self.max_initial_timestamp_index is None: _UpperCamelCase = model_config.vocab_size def __call__( self : int , lowercase_ : List[str] , lowercase_ : List[Any] , lowercase_ : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = scores.at[:, self.no_timestamps_token_id].set(-float("inf")) def handle_pairs(lowercase_ : Optional[Any] , lowercase_ : int): _UpperCamelCase = jnp.where((cur_len - self.begin_index) >= 1 , lowercase_ , lowercase_) _UpperCamelCase = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , lowercase_ , ) _UpperCamelCase = jnp.where((cur_len - self.begin_index) < 2 , lowercase_ , lowercase_) _UpperCamelCase = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , lowercase_ , lowercase_ , ) return jnp.where( lowercase_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float("inf")) , scores_k.at[: self.eos_token_id].set(-float("inf")) , ) , lowercase_ , ) _UpperCamelCase = jax.vmap(lowercase_)(lowercase_ , lowercase_) _UpperCamelCase = jnp.where(cur_len == self.begin_index , lowercase_ , lowercase_) _UpperCamelCase = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , lowercase_ , ) _UpperCamelCase = self.timestamp_begin + self.max_initial_timestamp_index _UpperCamelCase = jnp.where( lowercase_ , scores.at[:, last_allowed + 1 :].set(-float("inf")) , lowercase_ , ) # if sum of probability over timestamps is above any other token, sample timestamp _UpperCamelCase = jax.nn.log_softmax(lowercase_ , axis=-1) def handle_cumulative_probs(lowercase_ : int , lowercase_ : List[str]): _UpperCamelCase = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1) _UpperCamelCase = jnp.max(logprobs_k[: self.timestamp_begin]) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float("inf")) , lowercase_ , ) _UpperCamelCase = jax.vmap(lowercase_)(lowercase_ , lowercase_) return scores
709
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , *lowercase_ : List[str] , **lowercase_ : Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
82
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''edbeeching/decision-transformer-gym-hopper-medium''': ( '''https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json''' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''decision_transformer''' __A = ['''past_key_values'''] __A = { '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Dict , lowercase_ : Union[str, Any]=17 , lowercase_ : int=4 , lowercase_ : List[str]=128 , lowercase_ : List[Any]=4096 , lowercase_ : Dict=True , lowercase_ : Any=1 , lowercase_ : Optional[int]=1024 , lowercase_ : Optional[Any]=3 , lowercase_ : Optional[Any]=1 , lowercase_ : List[str]=None , lowercase_ : Any="relu" , lowercase_ : Tuple=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : Any=0.1 , lowercase_ : List[Any]=1e-5 , lowercase_ : Union[str, Any]=0.02 , lowercase_ : int=True , lowercase_ : Optional[int]=True , lowercase_ : List[Any]=50256 , lowercase_ : List[Any]=50256 , lowercase_ : Tuple=False , lowercase_ : Any=False , **lowercase_ : Union[str, Any] , ) -> int: """simple docstring""" _UpperCamelCase = state_dim _UpperCamelCase = act_dim _UpperCamelCase = hidden_size _UpperCamelCase = max_ep_len _UpperCamelCase = action_tanh _UpperCamelCase = vocab_size _UpperCamelCase = n_positions _UpperCamelCase = n_layer _UpperCamelCase = n_head _UpperCamelCase = n_inner _UpperCamelCase = activation_function _UpperCamelCase = resid_pdrop _UpperCamelCase = embd_pdrop _UpperCamelCase = attn_pdrop _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = scale_attn_weights _UpperCamelCase = use_cache _UpperCamelCase = scale_attn_by_inverse_layer_idx _UpperCamelCase = reorder_and_upcast_attn _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id super().__init__(bos_token_id=lowercase_ , eos_token_id=lowercase_ , **lowercase_)
710
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
82
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''google/vit-base-patch16-224''': '''https://huggingface.co/vit-base-patch16-224/resolve/main/config.json''', # See all ViT models at https://huggingface.co/models?filter=vit } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''vit''' def __init__( self : List[Any] , lowercase_ : Union[str, Any]=768 , lowercase_ : List[Any]=12 , lowercase_ : List[Any]=12 , lowercase_ : List[Any]=3072 , lowercase_ : Optional[Any]="gelu" , lowercase_ : List[str]=0.0 , lowercase_ : Union[str, Any]=0.0 , lowercase_ : Dict=0.02 , lowercase_ : List[str]=1e-1_2 , lowercase_ : Any=224 , lowercase_ : int=16 , lowercase_ : Optional[Any]=3 , lowercase_ : Dict=True , lowercase_ : int=16 , **lowercase_ : Optional[int] , ) -> Tuple: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = qkv_bias _UpperCamelCase = encoder_stride class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = version.parse('''1.11''' ) @property def __UpperCAmelCase ( self : Union[str, Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ]) @property def __UpperCAmelCase ( self : Dict) -> float: """simple docstring""" return 1e-4
711
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
82
0
import flax.linen as nn import jax import jax.numpy as jnp class _UpperCAmelCase ( nn.Module ): '''simple docstring''' __A = 42 __A = jnp.floataa def __UpperCAmelCase ( self : Tuple) -> int: """simple docstring""" _UpperCamelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Tuple , lowercase_ : str) -> str: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = hidden_states.shape _UpperCamelCase = jax.image.resize( lowercase_ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) _UpperCamelCase = self.conv(lowercase_) return hidden_states class _UpperCAmelCase ( nn.Module ): '''simple docstring''' __A = 42 __A = jnp.floataa def __UpperCAmelCase ( self : Dict) -> Optional[Any]: """simple docstring""" _UpperCamelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : List[Any] , lowercase_ : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = self.conv(lowercase_) return hidden_states class _UpperCAmelCase ( nn.Module ): '''simple docstring''' __A = 42 __A = None __A = 0.0 __A = None __A = jnp.floataa def __UpperCAmelCase ( self : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.in_channels if self.out_channels is None else self.out_channels _UpperCamelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5) _UpperCamelCase = nn.Conv( lowercase_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _UpperCamelCase = nn.Dense(lowercase_ , dtype=self.dtype) _UpperCamelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5) _UpperCamelCase = nn.Dropout(self.dropout_prob) _UpperCamelCase = nn.Conv( lowercase_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _UpperCamelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut _UpperCamelCase = None if use_nin_shortcut: _UpperCamelCase = nn.Conv( lowercase_ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] , lowercase_ : str=True) -> List[Any]: """simple docstring""" _UpperCamelCase = hidden_states _UpperCamelCase = self.norma(lowercase_) _UpperCamelCase = nn.swish(lowercase_) _UpperCamelCase = self.conva(lowercase_) _UpperCamelCase = self.time_emb_proj(nn.swish(lowercase_)) _UpperCamelCase = jnp.expand_dims(jnp.expand_dims(lowercase_ , 1) , 1) _UpperCamelCase = hidden_states + temb _UpperCamelCase = self.norma(lowercase_) _UpperCamelCase = nn.swish(lowercase_) _UpperCamelCase = self.dropout(lowercase_ , lowercase_) _UpperCamelCase = self.conva(lowercase_) if self.conv_shortcut is not None: _UpperCamelCase = self.conv_shortcut(lowercase_) return hidden_states + residual
712
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
82
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = 42 __A = 42 __A = None class _UpperCAmelCase ( lowerCAmelCase, lowerCAmelCase ): '''simple docstring''' __A = 2 @register_to_config def __init__( self : Tuple , lowercase_ : float = 0.02 , lowercase_ : float = 100 , lowercase_ : float = 1.0_07 , lowercase_ : float = 80 , lowercase_ : float = 0.05 , lowercase_ : float = 50 , ) -> str: """simple docstring""" _UpperCamelCase = sigma_max # setable values _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None # sigma(t_i) def __UpperCAmelCase ( self : Any , lowercase_ : torch.FloatTensor , lowercase_ : Optional[int] = None) -> torch.FloatTensor: """simple docstring""" return sample def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , lowercase_ : Union[str, torch.device] = None) -> Dict: """simple docstring""" _UpperCamelCase = num_inference_steps _UpperCamelCase = np.arange(0 , self.num_inference_steps)[::-1].copy() _UpperCamelCase = torch.from_numpy(lowercase_).to(lowercase_) _UpperCamelCase = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] _UpperCamelCase = torch.tensor(lowercase_ , dtype=torch.floataa , device=lowercase_) def __UpperCAmelCase ( self : Dict , lowercase_ : torch.FloatTensor , lowercase_ : float , lowercase_ : Optional[torch.Generator] = None) -> Tuple[torch.FloatTensor, float]: """simple docstring""" if self.config.s_min <= sigma <= self.config.s_max: _UpperCamelCase = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1) else: _UpperCamelCase = 0 # sample eps ~ N(0, S_noise^2 * I) _UpperCamelCase = self.config.s_noise * randn_tensor(sample.shape , generator=lowercase_).to(sample.device) _UpperCamelCase = sigma + gamma * sigma _UpperCamelCase = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __UpperCAmelCase ( self : str , lowercase_ : torch.FloatTensor , lowercase_ : float , lowercase_ : float , lowercase_ : torch.FloatTensor , lowercase_ : bool = True , ) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" _UpperCamelCase = sample_hat + sigma_hat * model_output _UpperCamelCase = (sample_hat - pred_original_sample) / sigma_hat _UpperCamelCase = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowercase_ , derivative=lowercase_ , pred_original_sample=lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : torch.FloatTensor , lowercase_ : float , lowercase_ : float , lowercase_ : torch.FloatTensor , lowercase_ : torch.FloatTensor , lowercase_ : torch.FloatTensor , lowercase_ : bool = True , ) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" _UpperCamelCase = sample_prev + sigma_prev * model_output _UpperCamelCase = (sample_prev - pred_original_sample) / sigma_prev _UpperCamelCase = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowercase_ , derivative=lowercase_ , pred_original_sample=lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Any) -> Any: """simple docstring""" raise NotImplementedError()
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase : '''simple docstring''' def __init__( self : str , lowercase_ : Tuple , lowercase_ : Dict=2 , lowercase_ : int=8 , lowercase_ : Optional[Any]=True , lowercase_ : Union[str, Any]=True , lowercase_ : List[Any]=True , lowercase_ : Union[str, Any]=True , lowercase_ : Dict=99 , lowercase_ : str=16 , lowercase_ : Union[str, Any]=5 , lowercase_ : Optional[Any]=2 , lowercase_ : List[str]=36 , lowercase_ : List[str]="gelu" , lowercase_ : List[Any]=0.0 , lowercase_ : int=0.0 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=16 , lowercase_ : Dict=2 , lowercase_ : List[str]=0.02 , lowercase_ : Union[str, Any]=3 , lowercase_ : List[str]=4 , lowercase_ : str=None , ) -> List[str]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def __UpperCAmelCase ( self : str) -> Optional[int]: """simple docstring""" _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self : Union[str, Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.get_config() _UpperCamelCase = 300 return config def __UpperCAmelCase ( self : List[str]) -> int: """simple docstring""" ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = self.prepare_config_and_inputs() _UpperCamelCase = True _UpperCamelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __UpperCAmelCase ( self : Dict , lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : str , lowercase_ : int , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : Dict) -> Any: """simple docstring""" _UpperCamelCase = MraModel(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_) _UpperCamelCase = model(lowercase_ , token_type_ids=lowercase_) _UpperCamelCase = model(lowercase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __UpperCAmelCase ( self : int , lowercase_ : str , lowercase_ : Tuple , lowercase_ : Dict , lowercase_ : Optional[int] , lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : str , lowercase_ : Tuple , ) -> List[Any]: """simple docstring""" _UpperCamelCase = True _UpperCamelCase = MraModel(lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , encoder_hidden_states=lowercase_ , encoder_attention_mask=lowercase_ , ) _UpperCamelCase = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , encoder_hidden_states=lowercase_ , ) _UpperCamelCase = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : int , lowercase_ : int , lowercase_ : Dict , lowercase_ : Optional[int]) -> Tuple: """simple docstring""" _UpperCamelCase = MraForMaskedLM(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : int) -> int: """simple docstring""" _UpperCamelCase = MraForQuestionAnswering(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , start_positions=lowercase_ , end_positions=lowercase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[Any] , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : str , lowercase_ : int , lowercase_ : Any , lowercase_ : int) -> str: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = MraForSequenceClassification(lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __UpperCAmelCase ( self : Tuple , lowercase_ : int , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Optional[int] , lowercase_ : Dict) -> Any: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = MraForTokenClassification(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def __UpperCAmelCase ( self : str , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : str , lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = self.num_choices _UpperCamelCase = MraForMultipleChoice(config=lowercase_) model.to(lowercase_) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __UpperCAmelCase ( self : Any) -> Optional[Any]: """simple docstring""" _UpperCamelCase = MraModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=lowercase_ , hidden_size=37) def __UpperCAmelCase ( self : Dict) -> Any: """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_) def __UpperCAmelCase ( self : Dict) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*lowercase_) def __UpperCAmelCase ( self : List[Any]) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase_) def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_) def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_) @slow def __UpperCAmelCase ( self : Optional[int]) -> Optional[int]: """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = MraModel.from_pretrained(lowercase_) self.assertIsNotNone(lowercase_) @unittest.skip(reason="MRA does not output attentions") def __UpperCAmelCase ( self : Optional[Any]) -> List[Any]: """simple docstring""" return @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = MraModel.from_pretrained("uw-madison/mra-base-512-4") _UpperCamelCase = torch.arange(256).unsqueeze(0) with torch.no_grad(): _UpperCamelCase = model(lowercase_)[0] _UpperCamelCase = torch.Size((1, 256, 768)) self.assertEqual(output.shape , lowercase_) _UpperCamelCase = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1e-4)) @slow def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" _UpperCamelCase = MraForMaskedLM.from_pretrained("uw-madison/mra-base-512-4") _UpperCamelCase = torch.arange(256).unsqueeze(0) with torch.no_grad(): _UpperCamelCase = model(lowercase_)[0] _UpperCamelCase = 50265 _UpperCamelCase = torch.Size((1, 256, vocab_size)) self.assertEqual(output.shape , lowercase_) _UpperCamelCase = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1e-4)) @slow def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = MraForMaskedLM.from_pretrained("uw-madison/mra-base-4096-8-d3") _UpperCamelCase = torch.arange(4096).unsqueeze(0) with torch.no_grad(): _UpperCamelCase = model(lowercase_)[0] _UpperCamelCase = 50265 _UpperCamelCase = torch.Size((1, 4096, vocab_size)) self.assertEqual(output.shape , lowercase_) _UpperCamelCase = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1e-4))
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = ['''pixel_values'''] def __init__( self : Tuple , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : bool = True , lowercase_ : Union[int, float] = 1 / 255 , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : bool = True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = size if size is not None else {"shortest_edge": 224} _UpperCamelCase = get_size_dict(lowercase_ , default_to_square=lowercase_) _UpperCamelCase = crop_size if crop_size is not None else {"height": 256, "width": 256} _UpperCamelCase = get_size_dict(lowercase_ , param_name="crop_size") _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = resample _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def __UpperCAmelCase ( self : Optional[int] , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : PILImageResampling = PIL.Image.BILINEAR , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Tuple , ) -> np.ndarray: """simple docstring""" _UpperCamelCase = get_size_dict(lowercase_ , default_to_square=lowercase_) if "shortest_edge" not in size: raise ValueError(f'The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}') _UpperCamelCase = get_resize_output_image_size(lowercase_ , size=size["shortest_edge"] , default_to_square=lowercase_) return resize(lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Union[str, Any] , ) -> np.ndarray: """simple docstring""" _UpperCamelCase = get_size_dict(lowercase_) if "height" not in size or "width" not in size: raise ValueError(f'The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}') return center_crop(lowercase_ , size=(size["height"], size["width"]) , data_format=lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Dict , lowercase_ : np.ndarray , lowercase_ : Union[int, float] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Optional[int] , ) -> int: """simple docstring""" return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_) def __UpperCAmelCase ( self : str , lowercase_ : np.ndarray , lowercase_ : Optional[Union[str, ChannelDimension]] = None) -> np.ndarray: """simple docstring""" return flip_channel_order(lowercase_ , data_format=lowercase_) def __UpperCAmelCase ( self : str , lowercase_ : ImageInput , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = None , lowercase_ : bool = None , lowercase_ : float = None , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : bool = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : ChannelDimension = ChannelDimension.FIRST , **lowercase_ : Optional[Any] , ) -> PIL.Image.Image: """simple docstring""" _UpperCamelCase = do_resize if do_resize is not None else self.do_resize _UpperCamelCase = resample if resample is not None else self.resample _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_center_crop if do_center_crop is not None else self.do_center_crop _UpperCamelCase = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) _UpperCamelCase = size if size is not None else self.size _UpperCamelCase = get_size_dict(lowercase_ , default_to_square=lowercase_) _UpperCamelCase = crop_size if crop_size is not None else self.crop_size _UpperCamelCase = get_size_dict(lowercase_ , param_name="crop_size") _UpperCamelCase = make_list_of_images(lowercase_) if not valid_images(lowercase_): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True.") # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(lowercase_) for image in images] if do_resize: _UpperCamelCase = [self.resize(image=lowercase_ , size=lowercase_ , resample=lowercase_) for image in images] if do_center_crop: _UpperCamelCase = [self.center_crop(image=lowercase_ , size=lowercase_) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=lowercase_ , scale=lowercase_) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: _UpperCamelCase = [self.flip_channel_order(image=lowercase_) for image in images] _UpperCamelCase = [to_channel_dimension_format(lowercase_ , lowercase_) for image in images] _UpperCamelCase = {"pixel_values": images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_) def __UpperCAmelCase ( self : Any , lowercase_ : Optional[int] , lowercase_ : List[Tuple] = None) -> Optional[Any]: """simple docstring""" _UpperCamelCase = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowercase_) != len(lowercase_): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits") if is_torch_tensor(lowercase_): _UpperCamelCase = target_sizes.numpy() _UpperCamelCase = [] for idx in range(len(lowercase_)): _UpperCamelCase = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0) , size=target_sizes[idx] , mode="bilinear" , align_corners=lowercase_) _UpperCamelCase = resized_logits[0].argmax(dim=0) semantic_segmentation.append(lowercase_) else: _UpperCamelCase = logits.argmax(dim=1) _UpperCamelCase = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0])] return semantic_segmentation
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase__ ( a__ ) ->Optional[int]: '''simple docstring''' return EnvironmentCommand() class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' @staticmethod def __UpperCAmelCase ( lowercase_ : ArgumentParser) -> Optional[Any]: """simple docstring""" _UpperCamelCase = parser.add_parser("env") download_parser.set_defaults(func=lowercase_) def __UpperCAmelCase ( self : Any) -> List[Any]: """simple docstring""" _UpperCamelCase = huggingface_hub.__version__ _UpperCamelCase = "not installed" _UpperCamelCase = "NA" if is_torch_available(): import torch _UpperCamelCase = torch.__version__ _UpperCamelCase = torch.cuda.is_available() _UpperCamelCase = "not installed" if is_transformers_available(): import transformers _UpperCamelCase = transformers.__version__ _UpperCamelCase = "not installed" if is_accelerate_available(): import accelerate _UpperCamelCase = accelerate.__version__ _UpperCamelCase = "not installed" if is_xformers_available(): import xformers _UpperCamelCase = xformers.__version__ _UpperCamelCase = { "`diffusers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "PyTorch version (GPU?)": f'{pt_version} ({pt_cuda_available})', "Huggingface_hub version": hub_version, "Transformers version": transformers_version, "Accelerate version": accelerate_version, "xFormers version": xformers_version, "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n") print(self.format_dict(lowercase_)) return info @staticmethod def __UpperCAmelCase ( lowercase_ : str) -> Optional[int]: """simple docstring""" return "\n".join([f'- {prop}: {val}' for prop, val in d.items()]) + "\n"
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask lowerCamelCase__ = logging.getLogger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[int] , lowercase_ : Any=-1) -> List[Any]: """simple docstring""" _UpperCamelCase = label_idx def __UpperCAmelCase ( self : str , lowercase_ : Tuple , lowercase_ : Union[Split, str]) -> List[InputExample]: """simple docstring""" if isinstance(lowercase_ , lowercase_): _UpperCamelCase = mode.value _UpperCamelCase = os.path.join(lowercase_ , f'{mode}.txt') _UpperCamelCase = 1 _UpperCamelCase = [] with open(lowercase_ , encoding="utf-8") as f: _UpperCamelCase = [] _UpperCamelCase = [] for line in f: if line.startswith("-DOCSTART-") or line == "" or line == "\n": if words: examples.append(InputExample(guid=f'{mode}-{guid_index}' , words=lowercase_ , labels=lowercase_)) guid_index += 1 _UpperCamelCase = [] _UpperCamelCase = [] else: _UpperCamelCase = line.split(" ") words.append(splits[0]) if len(lowercase_) > 1: labels.append(splits[self.label_idx].replace("\n" , "")) else: # Examples could have no label for mode = "test" labels.append("O") if words: examples.append(InputExample(guid=f'{mode}-{guid_index}' , words=lowercase_ , labels=lowercase_)) return examples def __UpperCAmelCase ( self : Any , lowercase_ : TextIO , lowercase_ : TextIO , lowercase_ : List) -> int: """simple docstring""" _UpperCamelCase = 0 for line in test_input_reader: if line.startswith("-DOCSTART-") or line == "" or line == "\n": writer.write(lowercase_) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _UpperCamelCase = line.split()[0] + " " + preds_list[example_id].pop(0) + "\n" writer.write(lowercase_) else: logger.warning("Maximum sequence length exceeded: No prediction for '%s'." , line.split()[0]) def __UpperCAmelCase ( self : Tuple , lowercase_ : str) -> List[str]: """simple docstring""" if path: with open(lowercase_ , "r") as f: _UpperCamelCase = f.read().splitlines() if "O" not in labels: _UpperCamelCase = ["O"] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any]) -> int: """simple docstring""" super().__init__(label_idx=-2) def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> List[str]: """simple docstring""" if path: with open(lowercase_ , "r") as f: _UpperCamelCase = f.read().splitlines() if "O" not in labels: _UpperCamelCase = ["O"] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple , lowercase_ : List[str] , lowercase_ : Union[Split, str]) -> List[InputExample]: """simple docstring""" if isinstance(lowercase_ , lowercase_): _UpperCamelCase = mode.value _UpperCamelCase = os.path.join(lowercase_ , f'{mode}.txt') _UpperCamelCase = 1 _UpperCamelCase = [] with open(lowercase_ , encoding="utf-8") as f: for sentence in parse_incr(lowercase_): _UpperCamelCase = [] _UpperCamelCase = [] for token in sentence: words.append(token["form"]) labels.append(token["upos"]) assert len(lowercase_) == len(lowercase_) if words: examples.append(InputExample(guid=f'{mode}-{guid_index}' , words=lowercase_ , labels=lowercase_)) guid_index += 1 return examples def __UpperCAmelCase ( self : Dict , lowercase_ : TextIO , lowercase_ : TextIO , lowercase_ : List) -> Tuple: """simple docstring""" _UpperCamelCase = 0 for sentence in parse_incr(lowercase_): _UpperCamelCase = preds_list[example_id] _UpperCamelCase = "" for token in sentence: out += f'{token["form"]} ({token["upos"]}|{s_p.pop(0)}) ' out += "\n" writer.write(lowercase_) example_id += 1 def __UpperCAmelCase ( self : Any , lowercase_ : str) -> List[str]: """simple docstring""" if path: with open(lowercase_ , "r") as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
def lowerCAmelCase__ ( a__ , a__ , a__ ) ->float: '''simple docstring''' return round(float(moles / volume ) * nfactor ) def lowerCAmelCase__ ( a__ , a__ , a__ ) ->float: '''simple docstring''' return round(float((moles * 0.0821 * temperature) / (volume) ) ) def lowerCAmelCase__ ( a__ , a__ , a__ ) ->float: '''simple docstring''' return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def lowerCAmelCase__ ( a__ , a__ , a__ ) ->float: '''simple docstring''' return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
def lowerCAmelCase__ ( a__ ) ->float: '''simple docstring''' if edge <= 0 or not isinstance(a__ , a__ ): raise ValueError("Length must be a positive." ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def lowerCAmelCase__ ( a__ ) ->float: '''simple docstring''' if edge <= 0 or not isinstance(a__ , a__ ): raise ValueError("Length must be a positive." ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
def lowerCAmelCase__ ( a__ ) ->bool: '''simple docstring''' _UpperCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack _UpperCamelCase = set() return any( node not in visited and depth_first_search(a__ , a__ , a__ , a__ ) for node in graph ) def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->bool: '''simple docstring''' visited.add(a__ ) rec_stk.add(a__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(a__ , a__ , a__ , a__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(a__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
700
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=lowercase_ , output_type="numpy" , return_dict=lowercase_)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = "google/ncsnpp-celebahq-256" _UpperCamelCase = UNetaDModel.from_pretrained(lowercase_) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=lowercase_ , scheduler=lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=lowercase_ , output_type="numpy").images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _UpperCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
82
0
lowerCamelCase__ = 9.80665 def lowerCAmelCase__ ( a__ , a__ , a__ = g ) ->float: '''simple docstring''' if fluid_density <= 0: raise ValueError("Impossible fluid density" ) if volume < 0: raise ValueError("Impossible Object volume" ) if gravity <= 0: raise ValueError("Impossible Gravity" ) return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
701
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __A = MODEL_FOR_MASKED_LM_MAPPING __A = TF_MODEL_FOR_MASKED_LM_MAPPING def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __UpperCAmelCase ( self : Tuple) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-0_5, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-0_5, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-0_5, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt") _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-0_5, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ {"sequence": "My name is Patrick", "score": 2.1e-0_5, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-0_5, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-0_5, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCamelCase = unmasker("My name is <mask> <mask>" , top_k=2) self.assertEqual( nested_simplify(lowercase_ , decimals=6) , [ [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-0_5, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-0_5, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt") # convert model to fp16 pipe.model.half() _UpperCamelCase = pipe("Paris is the [MASK] of France.") # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_) @slow @require_torch def __UpperCAmelCase ( self : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt") self.run_large_test(lowercase_) @slow @require_tf def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf") self.run_large_test(lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : int) -> Any: """simple docstring""" _UpperCamelCase = unmasker("My name is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is John", "score": 0.0_08, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_07, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCamelCase = unmasker("The largest city in France is <mask>") self.assertEqual( nested_simplify(lowercase_) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_51, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_14, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCamelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3) self.assertEqual( nested_simplify(lowercase_) , [ {"sequence": "My name is Patrick", "score": 0.0_05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_00, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_00, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) @require_tf def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf") _UpperCamelCase = None _UpperCamelCase = None self.run_pipeline_test(lowercase_ , []) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Optional[int]) -> int: """simple docstring""" if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)") _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = [ f'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = fill_masker.tokenizer _UpperCamelCase = fill_masker.model _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = fill_masker([f'This is a {tokenizer.mask_token}', f'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , ) with self.assertRaises(lowercase_): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase_): fill_masker("This is") self.run_test_top_k(lowercase_ , lowercase_) self.run_test_targets(lowercase_ , lowercase_) self.run_test_top_k_targets(lowercase_ , lowercase_) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int , lowercase_ : Dict , lowercase_ : List[str]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = sorted(vocab.keys())[:2] # Pipeline argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Call argument _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_) _UpperCamelCase = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_)) # Score equivalence _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCamelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_) == set(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=lowercase_) _UpperCamelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) # Raises with invalid with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets=[""]) with self.assertRaises(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , targets="") def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : str , lowercase_ : List[str]) -> Any: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}') self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ] , ) self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = tokenizer.get_vocab() _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) # top_k=2, ntargets=3 _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase_) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCamelCase = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_).issubset(lowercase_): _UpperCamelCase = fill_masker(f'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase_) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_) , nested_simplify(lowercase_)) def __UpperCAmelCase ( self : int , lowercase_ : Optional[int] , lowercase_ : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCamelCase = sorted(vocab.keys())[:3] _UpperCamelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCamelCase = fill_masker(f'My name is {tokenizer.mask_token}' , targets=lowercase_ , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_) , 3) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Any) -> Dict: """simple docstring""" _UpperCamelCase = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_) _UpperCamelCase = fill_masker( f'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], [ {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, {"sequence": ANY(lowercase_), "score": ANY(lowercase_), "token": ANY(lowercase_), "token_str": ANY(lowercase_)}, ], ] , )
82
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[int]) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __UpperCAmelCase ( self : str) -> List[str]: """simple docstring""" _UpperCamelCase = 1 _UpperCamelCase = 3 _UpperCamelCase = (32, 32) _UpperCamelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(lowercase_) return image @property def __UpperCAmelCase ( self : Tuple) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def __UpperCAmelCase ( self : List[str]) -> str: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def __UpperCAmelCase ( self : Tuple) -> Any: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(lowercase_) @property def __UpperCAmelCase ( self : str) -> str: """simple docstring""" def extract(*lowercase_ : int , **lowercase_ : str): class _UpperCAmelCase : '''simple docstring''' def __init__( self : int) -> Any: """simple docstring""" _UpperCamelCase = torch.ones([0]) def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" self.pixel_values.to(lowercase_) return self return Out() return extract def __UpperCAmelCase ( self : str) -> int: """simple docstring""" _UpperCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.dummy_cond_unet _UpperCamelCase = PNDMScheduler(skip_prk_steps=lowercase_) _UpperCamelCase = self.dummy_vae _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta") _UpperCamelCase = 77 _UpperCamelCase = self.dummy_image.to(lowercase_) _UpperCamelCase = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk _UpperCamelCase = AltDiffusionImgaImgPipeline( unet=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , safety_checker=lowercase_ , feature_extractor=self.dummy_extractor , ) _UpperCamelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase_) _UpperCamelCase = alt_pipe.to(lowercase_) alt_pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(0) _UpperCamelCase = alt_pipe( [prompt] , generator=lowercase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowercase_ , ) _UpperCamelCase = output.images _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(0) _UpperCamelCase = alt_pipe( [prompt] , generator=lowercase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowercase_ , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.44_27, 0.37_31, 0.42_49, 0.49_41, 0.45_46, 0.41_48, 0.41_93, 0.46_66, 0.44_99]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 5e-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU") def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" _UpperCamelCase = self.dummy_cond_unet _UpperCamelCase = PNDMScheduler(skip_prk_steps=lowercase_) _UpperCamelCase = self.dummy_vae _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta") _UpperCamelCase = 77 _UpperCamelCase = self.dummy_image.to(lowercase_) # put models in fp16 _UpperCamelCase = unet.half() _UpperCamelCase = vae.half() _UpperCamelCase = bert.half() # make sure here that pndm scheduler skips prk _UpperCamelCase = AltDiffusionImgaImgPipeline( unet=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , safety_checker=lowercase_ , feature_extractor=self.dummy_extractor , ) _UpperCamelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase_) _UpperCamelCase = alt_pipe.to(lowercase_) alt_pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = alt_pipe( [prompt] , generator=lowercase_ , num_inference_steps=2 , output_type="np" , image=lowercase_ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU") def __UpperCAmelCase ( self : Optional[Any]) -> str: """simple docstring""" _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg") # resize to resolution that is divisible by 8 but not 16 or 32 _UpperCamelCase = init_image.resize((760, 504)) _UpperCamelCase = "BAAI/AltDiffusion" _UpperCamelCase = AltDiffusionImgaImgPipeline.from_pretrained( lowercase_ , safety_checker=lowercase_ , ) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) pipe.enable_attention_slicing() _UpperCamelCase = "A fantasy landscape, trending on artstation" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , image=lowercase_ , strength=0.75 , guidance_scale=7.5 , generator=lowercase_ , output_type="np" , ) _UpperCamelCase = output.images[0] _UpperCamelCase = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) _UpperCamelCase = np.array([0.93_58, 0.93_97, 0.95_99, 0.99_01, 1.00_00, 1.00_00, 0.98_82, 1.00_00, 1.00_00]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg") _UpperCamelCase = init_image.resize((768, 512)) _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy") _UpperCamelCase = "BAAI/AltDiffusion" _UpperCamelCase = AltDiffusionImgaImgPipeline.from_pretrained( lowercase_ , safety_checker=lowercase_ , ) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) pipe.enable_attention_slicing() _UpperCamelCase = "A fantasy landscape, trending on artstation" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , image=lowercase_ , strength=0.75 , guidance_scale=7.5 , generator=lowercase_ , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image).max() < 1e-2
702
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCamelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a__ ) ) elif i in operators: # RULE 2 operator_stack.push(a__ ) elif i == ")": # RULE 4 _UpperCamelCase = operator_stack.peek() operator_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operand_stack.peek() operand_stack.pop() _UpperCamelCase = operators[opr](a__ , a__ ) operand_stack.push(a__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
0