code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
def lowerCAmelCase__ ( a__ ) ->Tuple: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = len(a__ ) for i in range(n - 1 ): for j in range(i + 1 , a__ ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' if len(a__ ) <= 1: return arr, 0 _UpperCamelCase = len(a__ ) // 2 _UpperCamelCase = arr[0:mid] _UpperCamelCase = arr[mid:] _UpperCamelCase , _UpperCamelCase = count_inversions_recursive(a__ ) _UpperCamelCase , _UpperCamelCase = count_inversions_recursive(a__ ) _UpperCamelCase , _UpperCamelCase = _count_cross_inversions(a__ , a__ ) _UpperCamelCase = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowerCAmelCase__ ( a__ , a__ ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = _UpperCamelCase = _UpperCamelCase = 0 while i < len(a__ ) and j < len(a__ ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(a__ ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(a__ ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowerCAmelCase__ ( ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) _UpperCamelCase = count_inversions_bf(a__ ) _UpperCamelCase , _UpperCamelCase = count_inversions_recursive(a__ ) assert num_inversions_bf == num_inversions_recursive == 8 print("number of inversions = " , a__ ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() _UpperCamelCase = count_inversions_bf(a__ ) _UpperCamelCase , _UpperCamelCase = count_inversions_recursive(a__ ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , a__ ) # an empty list should also have zero inversions _UpperCamelCase = [] _UpperCamelCase = count_inversions_bf(a__ ) _UpperCamelCase , _UpperCamelCase = count_inversions_recursive(a__ ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , a__ ) if __name__ == "__main__": main()
703
import logging from transformers import PretrainedConfig lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''bertabs''' def __init__( self : List[str] , lowercase_ : int=30522 , lowercase_ : str=512 , lowercase_ : int=6 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[Any]=8 , lowercase_ : Optional[int]=512 , lowercase_ : Tuple=0.2 , lowercase_ : Union[str, Any]=6 , lowercase_ : List[Any]=768 , lowercase_ : List[str]=8 , lowercase_ : int=2048 , lowercase_ : Tuple=0.2 , **lowercase_ : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
82
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/config.json''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/config.json''', '''funnel-transformer/medium-base''': '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json''', '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/config.json''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json''', '''funnel-transformer/xlarge-base''': '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''funnel''' __A = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', } def __init__( self : str , lowercase_ : Optional[Any]=30522 , lowercase_ : Union[str, Any]=[4, 4, 4] , lowercase_ : int=None , lowercase_ : List[str]=2 , lowercase_ : Optional[Any]=768 , lowercase_ : Tuple=12 , lowercase_ : int=64 , lowercase_ : List[Any]=3072 , lowercase_ : Optional[Any]="gelu_new" , lowercase_ : Optional[int]=0.1 , lowercase_ : Optional[int]=0.1 , lowercase_ : str=0.0 , lowercase_ : Dict=0.1 , lowercase_ : str=None , lowercase_ : Dict=1e-9 , lowercase_ : List[str]="mean" , lowercase_ : int="relative_shift" , lowercase_ : Tuple=True , lowercase_ : Dict=True , lowercase_ : Any=True , **lowercase_ : Union[str, Any] , ) -> Optional[int]: """simple docstring""" _UpperCamelCase = vocab_size _UpperCamelCase = block_sizes _UpperCamelCase = [1] * len(lowercase_) if block_repeats is None else block_repeats assert len(lowercase_) == len( self.block_repeats), "`block_sizes` and `block_repeats` should have the same length." _UpperCamelCase = num_decoder_layers _UpperCamelCase = d_model _UpperCamelCase = n_head _UpperCamelCase = d_head _UpperCamelCase = d_inner _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = initializer_range _UpperCamelCase = initializer_std _UpperCamelCase = layer_norm_eps assert pooling_type in [ "mean", "max", ], f'Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.' _UpperCamelCase = pooling_type assert attention_type in [ "relative_shift", "factorized", ], f'Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.' _UpperCamelCase = attention_type _UpperCamelCase = separate_cls _UpperCamelCase = truncate_seq _UpperCamelCase = pool_q_only super().__init__(**lowercase_) @property def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return sum(self.block_sizes) @num_hidden_layers.setter def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[Any]) -> str: """simple docstring""" raise NotImplementedError( "This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.") @property def __UpperCAmelCase ( self : Any) -> Dict: """simple docstring""" return len(self.block_sizes) @num_blocks.setter def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[str]) -> List[Any]: """simple docstring""" raise NotImplementedError("This model does not support the setting of `num_blocks`. Please set `block_sizes`.")
704
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase__ = input('''Enter image url: ''').strip() print(F"Downloading image from {url} ...") lowerCamelCase__ = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase__ = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCamelCase__ = requests.get(image_url).content lowerCamelCase__ = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
0
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCamelCase__ = numpy.array([0, 0]) lowerCamelCase__ = numpy.array([0.5, 0.8660254]) lowerCamelCase__ = numpy.array([1, 0]) lowerCamelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase__ ( a__ , a__ ) ->list[numpy.ndarray]: '''simple docstring''' _UpperCamelCase = initial_vectors for _ in range(a__ ): _UpperCamelCase = iteration_step(a__ ) return vectors def lowerCAmelCase__ ( a__ ) ->list[numpy.ndarray]: '''simple docstring''' _UpperCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _UpperCamelCase = vectors[i + 1] new_vectors.append(a__ ) _UpperCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase__ ( a__ , a__ ) ->numpy.ndarray: '''simple docstring''' _UpperCamelCase = numpy.radians(a__ ) _UpperCamelCase , _UpperCamelCase = numpy.cos(a__ ), numpy.sin(a__ ) _UpperCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(a__ , a__ ) def lowerCAmelCase__ ( a__ ) ->None: '''simple docstring''' _UpperCamelCase = plt.gca() axes.set_aspect("equal" ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _UpperCamelCase , _UpperCamelCase = zip(*a__ ) plt.plot(a__ , a__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
705
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''dpr''' def __init__( self : Optional[Any] , lowercase_ : int=30522 , lowercase_ : str=768 , lowercase_ : List[Any]=12 , lowercase_ : Dict=12 , lowercase_ : str=3072 , lowercase_ : Any="gelu" , lowercase_ : Any=0.1 , lowercase_ : Any=0.1 , lowercase_ : str=512 , lowercase_ : str=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Dict=1e-1_2 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]="absolute" , lowercase_ : int = 0 , **lowercase_ : int , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase_ , **lowercase_) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = projection_dim _UpperCamelCase = position_embedding_type
82
0
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch lowerCamelCase__ = '''sshleifer/bart-tiny-random''' lowerCamelCase__ = '''patrickvonplaten/t5-tiny-random''' @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def __UpperCAmelCase ( self : List[Any]) -> Any: """simple docstring""" return AutoConfig.from_pretrained(lowercase_) def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" _UpperCamelCase , *_UpperCamelCase = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=1) self.assertEqual(student.config.num_hidden_layers , 1) def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase , *_UpperCamelCase = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=lowercase_) def __UpperCAmelCase ( self : Optional[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , *_UpperCamelCase = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=lowercase_) self.assertEqual(student.config.encoder_layers , 1) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers) def __UpperCAmelCase ( self : List[str]) -> List[str]: """simple docstring""" _UpperCamelCase , *_UpperCamelCase = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=1) self.assertEqual(student.config.encoder_layers , 1) self.assertEqual(student.config.decoder_layers , 1) def __UpperCAmelCase ( self : List[Any]) -> int: """simple docstring""" with self.assertRaises(lowercase_): create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=lowercase_ , d=lowercase_)
706
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
82
0
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCamelCase__ = False class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' pass @nightly @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[Any]) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[Any]) -> Dict: """simple docstring""" _UpperCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion") # remove text_unet pipe.remove_unused_weights() pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger " _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy").images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowercase_) _UpperCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = generator.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy").images assert np.abs(image - new_image).sum() < 1e-5, "Models don't have the same forward pass" def __UpperCAmelCase ( self : Dict) -> Any: """simple docstring""" _UpperCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger " _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy").images _UpperCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
707
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def lowerCAmelCase__ ( a__ , a__ , a__ , a__=None ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = XLNetConfig.from_json_file(a__ ) _UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) _UpperCamelCase = finetuning_task _UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCamelCase = XLNetForSequenceClassification(a__ ) elif "squad" in finetuning_task: _UpperCamelCase = finetuning_task _UpperCamelCase = XLNetForQuestionAnswering(a__ ) else: _UpperCamelCase = XLNetLMHeadModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a__ , a__ , a__ ) # Save pytorch-model _UpperCamelCase = os.path.join(a__ , a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) print(f'Save PyTorch model to {os.path.abspath(a__ )}' ) torch.save(model.state_dict() , a__ ) print(f'Save configuration file to {os.path.abspath(a__ )}' ) with open(a__ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
82
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[Any]) -> Dict: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __UpperCAmelCase ( self : Any) -> List[str]: """simple docstring""" _UpperCamelCase = 1 _UpperCamelCase = 3 _UpperCamelCase = (32, 32) _UpperCamelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(lowercase_) return image @property def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=lowercase_ , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def __UpperCAmelCase ( self : Dict) -> Tuple: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) return CLIPTextModel(lowercase_) def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" _UpperCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.dummy_cond_unet_upscale _UpperCamelCase = DDPMScheduler() _UpperCamelCase = DDIMScheduler(prediction_type="v_prediction") _UpperCamelCase = self.dummy_vae _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") _UpperCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((64, 64)) # make sure here that pndm scheduler skips prk _UpperCamelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=350 , ) _UpperCamelCase = sd_pipe.to(lowercase_) sd_pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(0) _UpperCamelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) _UpperCamelCase = output.images _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(0) _UpperCamelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] _UpperCamelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) _UpperCamelCase = np.array([0.31_13, 0.39_10, 0.42_72, 0.48_59, 0.50_61, 0.46_52, 0.53_62, 0.57_15, 0.56_61]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def __UpperCAmelCase ( self : Any) -> Any: """simple docstring""" _UpperCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.dummy_cond_unet_upscale _UpperCamelCase = DDPMScheduler() _UpperCamelCase = DDIMScheduler(prediction_type="v_prediction") _UpperCamelCase = self.dummy_vae _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") _UpperCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((64, 64)) # make sure here that pndm scheduler skips prk _UpperCamelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=350 , ) _UpperCamelCase = sd_pipe.to(lowercase_) sd_pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) _UpperCamelCase = output.images assert image.shape[0] == 2 _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(0) _UpperCamelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) _UpperCamelCase = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU") def __UpperCAmelCase ( self : Tuple) -> Dict: """simple docstring""" _UpperCamelCase = self.dummy_cond_unet_upscale _UpperCamelCase = DDPMScheduler() _UpperCamelCase = DDIMScheduler(prediction_type="v_prediction") _UpperCamelCase = self.dummy_vae _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") _UpperCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((64, 64)) # put models in fp16, except vae as it overflows in fp16 _UpperCamelCase = unet.half() _UpperCamelCase = text_encoder.half() # make sure here that pndm scheduler skips prk _UpperCamelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=350 , ) _UpperCamelCase = sd_pipe.to(lowercase_) sd_pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , num_inference_steps=2 , output_type="np" , ).images _UpperCamelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : int) -> List[Any]: """simple docstring""" _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat.npy") _UpperCamelCase = "stabilityai/stable-diffusion-x4-upscaler" _UpperCamelCase = StableDiffusionUpscalePipeline.from_pretrained(lowercase_) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) pipe.enable_attention_slicing() _UpperCamelCase = "a cat sitting on a park bench" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image).max() < 1e-3 def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat_fp16.npy") _UpperCamelCase = "stabilityai/stable-diffusion-x4-upscaler" _UpperCamelCase = StableDiffusionUpscalePipeline.from_pretrained( lowercase_ , torch_dtype=torch.floataa , ) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) pipe.enable_attention_slicing() _UpperCamelCase = "a cat sitting on a park bench" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image).max() < 5e-1 def __UpperCAmelCase ( self : Tuple) -> int: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png") _UpperCamelCase = "stabilityai/stable-diffusion-x4-upscaler" _UpperCamelCase = StableDiffusionUpscalePipeline.from_pretrained( lowercase_ , torch_dtype=torch.floataa , ) pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() _UpperCamelCase = "a cat sitting on a park bench" _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , num_inference_steps=5 , output_type="np" , ) _UpperCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
708
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple) -> int: """simple docstring""" super().__init__() _UpperCamelCase = model _UpperCamelCase = 2 _UpperCamelCase = nn.Linear(self.model.config.hidden_size , self.num_labels) def __UpperCAmelCase ( self : Union[str, Any]) -> Any: """simple docstring""" pass def lowerCAmelCase__ ( a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = LongformerModel.from_pretrained(a__ ) _UpperCamelCase = LightningModel(a__ ) _UpperCamelCase = torch.load(a__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCamelCase = LongformerForQuestionAnswering.from_pretrained(a__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a__ ) print(f'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
82
0
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/config.json''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/config.json''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/config.json''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/config.json''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/config.json''', } class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''t5''' __A = ['''past_key_values'''] __A = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self : Any , lowercase_ : Union[str, Any]=32128 , lowercase_ : Optional[Any]=512 , lowercase_ : Optional[int]=64 , lowercase_ : Any=2048 , lowercase_ : List[Any]=6 , lowercase_ : str=None , lowercase_ : List[Any]=8 , lowercase_ : str=32 , lowercase_ : Dict=128 , lowercase_ : List[Any]=0.1 , lowercase_ : Optional[Any]=1e-6 , lowercase_ : str=1.0 , lowercase_ : Optional[int]="relu" , lowercase_ : Tuple=True , lowercase_ : str=True , lowercase_ : int=0 , lowercase_ : Optional[Any]=1 , **lowercase_ : Union[str, Any] , ) -> List[str]: """simple docstring""" _UpperCamelCase = vocab_size _UpperCamelCase = d_model _UpperCamelCase = d_kv _UpperCamelCase = d_ff _UpperCamelCase = num_layers _UpperCamelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _UpperCamelCase = num_heads _UpperCamelCase = relative_attention_num_buckets _UpperCamelCase = relative_attention_max_distance _UpperCamelCase = dropout_rate _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_factor _UpperCamelCase = feed_forward_proj _UpperCamelCase = use_cache _UpperCamelCase = self.feed_forward_proj.split("-") _UpperCamelCase = act_info[-1] _UpperCamelCase = act_info[0] == "gated" if len(lowercase_) > 1 and act_info[0] != "gated" or len(lowercase_) > 2: raise ValueError( f'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "'gated-gelu' or 'relu'") # for backwards compatibility if feed_forward_proj == "gated-gelu": _UpperCamelCase = "gelu_new" super().__init__( pad_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , **lowercase_ , ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' @property def __UpperCAmelCase ( self : int) -> Mapping[str, Mapping[int, str]]: """simple docstring""" _UpperCamelCase = { "input_ids": {0: "batch", 1: "encoder_sequence"}, "attention_mask": {0: "batch", 1: "encoder_sequence"}, } if self.use_past: _UpperCamelCase = "past_encoder_sequence + sequence" _UpperCamelCase = {0: "batch"} _UpperCamelCase = {0: "batch", 1: "past_decoder_sequence + sequence"} else: _UpperCamelCase = {0: "batch", 1: "decoder_sequence"} _UpperCamelCase = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowercase_ , direction="inputs") return common_inputs @property def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" return 13
709
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , *lowercase_ : List[str] , **lowercase_ : Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
82
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''facebook/bart-large-mnli''' __A = ( '''This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ''' '''should be the text to classify, and `labels`, which should be the list of labels to use for classification. ''' '''It returns the most likely label in the list of provided `labels` for the input text.''' ) __A = '''text_classifier''' __A = AutoTokenizer __A = AutoModelForSequenceClassification __A = ['''text''', ['''text''']] __A = ['''text'''] def __UpperCAmelCase ( self : Any) -> Dict: """simple docstring""" super().setup() _UpperCamelCase = self.model.config _UpperCamelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("entail"): _UpperCamelCase = int(lowercase_) if self.entailment_id == -1: raise ValueError("Could not determine the entailment ID from the model config, please pass it at init.") def __UpperCAmelCase ( self : Any , lowercase_ : List[Any] , lowercase_ : Tuple) -> Dict: """simple docstring""" _UpperCamelCase = labels return self.pre_processor( [text] * len(lowercase_) , [f'This example is {label}' for label in labels] , return_tensors="pt" , padding="max_length" , ) def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any]) -> str: """simple docstring""" _UpperCamelCase = outputs.logits _UpperCamelCase = torch.argmax(logits[:, 2]).item() return self._labels[label_id]
710
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( enum.Enum ): '''simple docstring''' __A = 0 __A = 1 @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''generated''' def __init__( self : Any , *lowercase_ : Dict , **lowercase_ : Tuple) -> List[Any]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[int]=None , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = {} if truncation is not None: _UpperCamelCase = truncation _UpperCamelCase = generate_kwargs _UpperCamelCase = {} if return_tensors is not None and return_type is None: _UpperCamelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _UpperCamelCase = return_type if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCamelCase = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) if len(lowercase_) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim.") _UpperCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : int , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> Any: """simple docstring""" return True def __UpperCAmelCase ( self : Dict , *lowercase_ : List[str] , lowercase_ : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowercase_): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input") _UpperCamelCase = ([prefix + arg for arg in args[0]],) _UpperCamelCase = True elif isinstance(args[0] , lowercase_): _UpperCamelCase = (prefix + args[0],) _UpperCamelCase = False else: raise ValueError( f' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`') _UpperCamelCase = self.tokenizer(*lowercase_ , padding=lowercase_ , truncation=lowercase_ , return_tensors=self.framework) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[Any] , *lowercase_ : Any , **lowercase_ : int) -> Dict: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) if ( isinstance(args[0] , lowercase_) and all(isinstance(lowercase_ , lowercase_) for el in args[0]) and all(len(lowercase_) == 1 for res in result) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Tuple , lowercase_ : Union[str, Any] , lowercase_ : str=TruncationStrategy.DO_NOT_TRUNCATE , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = self._parse_and_tokenize(lowercase_ , truncation=lowercase_ , **lowercase_) return inputs def __UpperCAmelCase ( self : str , lowercase_ : str , **lowercase_ : str) -> str: """simple docstring""" if self.framework == "pt": _UpperCamelCase , _UpperCamelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _UpperCamelCase , _UpperCamelCase = tf.shape(model_inputs["input_ids"]).numpy() _UpperCamelCase = generate_kwargs.get("min_length" , self.model.config.min_length) _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) self.check_inputs(lowercase_ , generate_kwargs["min_length"] , generate_kwargs["max_length"]) _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) _UpperCamelCase = output_ids.shape[0] if self.framework == "pt": _UpperCamelCase = output_ids.reshape(lowercase_ , out_b // in_b , *output_ids.shape[1:]) elif self.framework == "tf": _UpperCamelCase = tf.reshape(lowercase_ , (in_b, out_b // in_b, *output_ids.shape[1:])) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Dict , lowercase_ : str , lowercase_ : int=ReturnType.TEXT , lowercase_ : int=False) -> Tuple: """simple docstring""" _UpperCamelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _UpperCamelCase = {f'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: _UpperCamelCase = { f'{self.return_name}_text': self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) } records.append(lowercase_) return records @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''summary''' def __call__( self : Optional[Any] , *lowercase_ : int , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f'Your min_length={min_length} must be inferior than your max_length={max_length}.') if input_length < max_length: logger.warning( f'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})') @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = '''translation''' def __UpperCAmelCase ( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : int) -> int: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)") return True def __UpperCAmelCase ( self : Tuple , *lowercase_ : Any , lowercase_ : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , lowercase_ : Any=None , lowercase_ : Optional[Any]=None) -> List[str]: """simple docstring""" if getattr(self.tokenizer , "_build_translation_inputs" , lowercase_): return self.tokenizer._build_translation_inputs( *lowercase_ , return_tensors=self.framework , truncation=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_) else: return super()._parse_and_tokenize(*lowercase_ , truncation=lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : Dict=None , lowercase_ : str=None , **lowercase_ : List[Any]) -> List[Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = super()._sanitize_parameters(**lowercase_) if src_lang is not None: _UpperCamelCase = src_lang if tgt_lang is not None: _UpperCamelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _UpperCamelCase = kwargs.get("task" , self.task) _UpperCamelCase = task.split("_") if task and len(lowercase_) == 4: # translation, XX, to YY _UpperCamelCase = items[1] _UpperCamelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *lowercase_ : List[str] , **lowercase_ : str) -> Union[str, Any]: """simple docstring""" return super().__call__(*lowercase_ , **lowercase_)
82
0
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging lowerCamelCase__ = ( '''https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py''' ) lowerCamelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' _UpperCamelCase = "https://pypi.org/pypi/diffusers/json" _UpperCamelCase = json.loads(request.urlopen(a__ ).read() )["releases"].keys() return sorted(a__ , key=lambda a__ : version.Version(a__ ) ) def lowerCAmelCase__ ( ) ->Union[str, Any]: '''simple docstring''' if HF_MODULES_CACHE in sys.path: return sys.path.append(a__ ) os.makedirs(a__ , exist_ok=a__ ) _UpperCamelCase = Path(a__ ) / "__init__.py" if not init_path.exists(): init_path.touch() def lowerCAmelCase__ ( a__ ) ->Tuple: '''simple docstring''' init_hf_modules() _UpperCamelCase = Path(a__ ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(a__ , exist_ok=a__ ) _UpperCamelCase = dynamic_module_path / "__init__.py" if not init_path.exists(): init_path.touch() def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' with open(a__ , "r" , encoding="utf-8" ) as f: _UpperCamelCase = f.read() # Imports of the form `import .xxx` _UpperCamelCase = re.findall("^\s*import\s+\.(\S+)\s*$" , a__ , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall("^\s*from\s+\.(\S+)\s+import" , a__ , flags=re.MULTILINE ) # Unique-ify return list(set(a__ ) ) def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' _UpperCamelCase = False _UpperCamelCase = [module_file] _UpperCamelCase = [] # Let's recurse through all relative imports while not no_change: _UpperCamelCase = [] for f in files_to_check: new_imports.extend(get_relative_imports(a__ ) ) _UpperCamelCase = Path(a__ ).parent _UpperCamelCase = [str(module_path / m ) for m in new_imports] _UpperCamelCase = [f for f in new_import_files if f not in all_relative_imports] _UpperCamelCase = [f'{f}.py' for f in new_import_files] _UpperCamelCase = len(a__ ) == 0 all_relative_imports.extend(a__ ) return all_relative_imports def lowerCAmelCase__ ( a__ ) ->List[Any]: '''simple docstring''' with open(a__ , "r" , encoding="utf-8" ) as f: _UpperCamelCase = f.read() # Imports of the form `import xxx` _UpperCamelCase = re.findall("^\s*import\s+(\S+)\s*$" , a__ , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall("^\s*from\s+(\S+)\s+import" , a__ , flags=re.MULTILINE ) # Only keep the top-level module _UpperCamelCase = [imp.split("." )[0] for imp in imports if not imp.startswith("." )] # Unique-ify and test we got them all _UpperCamelCase = list(set(a__ ) ) _UpperCamelCase = [] for imp in imports: try: importlib.import_module(a__ ) except ImportError: missing_packages.append(a__ ) if len(a__ ) > 0: raise ImportError( "This modeling file requires the following packages that were not found in your environment: " f'{", ".join(a__ )}. Run `pip install {" ".join(a__ )}`' ) return get_relative_imports(a__ ) def lowerCAmelCase__ ( a__ , a__ ) ->Optional[int]: '''simple docstring''' _UpperCamelCase = module_path.replace(os.path.sep , "." ) _UpperCamelCase = importlib.import_module(a__ ) if class_name is None: return find_pipeline_class(a__ ) return getattr(a__ , a__ ) def lowerCAmelCase__ ( a__ ) ->Tuple: '''simple docstring''' from ..pipelines import DiffusionPipeline _UpperCamelCase = dict(inspect.getmembers(a__ , inspect.isclass ) ) _UpperCamelCase = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , a__ ) and cls.__module__.split("." )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f'Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:' f' {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in' f' {loaded_module}.' ) _UpperCamelCase = cls return pipeline_class def lowerCAmelCase__ ( a__ , a__ , a__ = None , a__ = False , a__ = False , a__ = None , a__ = None , a__ = None , a__ = False , ) ->Tuple: '''simple docstring''' _UpperCamelCase = str(a__ ) _UpperCamelCase = os.path.join(a__ , a__ ) if os.path.isfile(a__ ): _UpperCamelCase = module_file_or_url _UpperCamelCase = "local" elif pretrained_model_name_or_path.count("/" ) == 0: _UpperCamelCase = get_diffusers_versions() # cut ".dev0" _UpperCamelCase = "v" + ".".join(__version__.split("." )[:3] ) # retrieve github version that matches if revision is None: _UpperCamelCase = latest_version if latest_version[1:] in available_versions else "main" logger.info(f'Defaulting to latest_version: {revision}.' ) elif revision in available_versions: _UpperCamelCase = f'v{revision}' elif revision == "main": _UpperCamelCase = revision else: raise ValueError( f'`custom_revision`: {revision} does not exist. Please make sure to choose one of' f' {", ".join(available_versions + ["main"] )}.' ) # community pipeline on GitHub _UpperCamelCase = COMMUNITY_PIPELINES_URL.format(revision=a__ , pipeline=a__ ) try: _UpperCamelCase = cached_download( a__ , cache_dir=a__ , force_download=a__ , proxies=a__ , resume_download=a__ , local_files_only=a__ , use_auth_token=a__ , ) _UpperCamelCase = "git" _UpperCamelCase = pretrained_model_name_or_path + ".py" except EnvironmentError: logger.error(f'Could not locate the {module_file} inside {pretrained_model_name_or_path}.' ) raise else: try: # Load from URL or cache if already cached _UpperCamelCase = hf_hub_download( a__ , a__ , cache_dir=a__ , force_download=a__ , proxies=a__ , resume_download=a__ , local_files_only=a__ , use_auth_token=a__ , ) _UpperCamelCase = os.path.join("local" , "--".join(pretrained_model_name_or_path.split("/" ) ) ) except EnvironmentError: logger.error(f'Could not locate the {module_file} inside {pretrained_model_name_or_path}.' ) raise # Check we have all the requirements in our environment _UpperCamelCase = check_imports(a__ ) # Now we move the module inside our cached dynamic modules. _UpperCamelCase = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(a__ ) _UpperCamelCase = Path(a__ ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(a__ , submodule_path / module_file ) for module_needed in modules_needed: _UpperCamelCase = f'{module_needed}.py' shutil.copy(os.path.join(a__ , a__ ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(a__ , a__ ): _UpperCamelCase = use_auth_token elif use_auth_token is True: _UpperCamelCase = HfFolder.get_token() else: _UpperCamelCase = None _UpperCamelCase = model_info(a__ , revision=a__ , token=a__ ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _UpperCamelCase = submodule_path / commit_hash _UpperCamelCase = full_submodule + os.path.sep + commit_hash create_dynamic_module(a__ ) if not (submodule_path / module_file).exists(): shutil.copy(a__ , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( a__ , f'{module_needed}.py' , cache_dir=a__ , force_download=a__ , resume_download=a__ , proxies=a__ , use_auth_token=a__ , revision=a__ , local_files_only=a__ , ) return os.path.join(a__ , a__ ) def lowerCAmelCase__ ( a__ , a__ , a__ = None , a__ = None , a__ = False , a__ = False , a__ = None , a__ = None , a__ = None , a__ = False , **a__ , ) ->Tuple: '''simple docstring''' _UpperCamelCase = get_cached_module_file( a__ , a__ , cache_dir=a__ , force_download=a__ , resume_download=a__ , proxies=a__ , use_auth_token=a__ , revision=a__ , local_files_only=a__ , ) return get_class_in_module(a__ , final_module.replace(".py" , "" ) )
711
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCamelCase__ = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', } } # TODO(PVP) - this should be removed in Transformers v5 lowerCamelCase__ = { '''t5-small''': 512, '''t5-base''': 512, '''t5-large''': 512, '''t5-3b''': 512, '''t5-11b''': 512, } lowerCamelCase__ = '''▁''' class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ['''input_ids''', '''attention_mask'''] def __init__( self : Tuple , lowercase_ : int , lowercase_ : str="</s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : Dict="<pad>" , lowercase_ : Tuple=100 , lowercase_ : str=None , lowercase_ : Optional[Dict[str, Any]] = None , lowercase_ : str=True , **lowercase_ : Optional[Any] , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [f'<extra_id_{i}>' for i in range(lowercase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda lowercase_: bool("extra_id" in str(lowercase_)) , lowercase_))) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens") if legacy: logger.warning_once( f'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565") _UpperCamelCase = legacy _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , extra_ids=lowercase_ , additional_special_tokens=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=lowercase_ , **lowercase_ , ) _UpperCamelCase = vocab_file _UpperCamelCase = extra_ids _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowercase_) @staticmethod def __UpperCAmelCase ( lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : str) -> Any: """simple docstring""" if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _UpperCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowercase_ , ) return max_model_length @property def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() + self._extra_ids def __UpperCAmelCase ( self : Dict) -> Optional[int]: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase_)) + [1] return ([0] * len(lowercase_)) + [1] + ([0] * len(lowercase_)) + [1] def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" return list( set(filter(lambda lowercase_: bool(re.search(R"<extra_id_\d+>" , lowercase_)) is not None , self.additional_special_tokens))) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" return [self._convert_token_to_id(lowercase_) for token in self.get_sentinel_tokens()] def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : List[int]) -> List[int]: """simple docstring""" if len(lowercase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( f'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def __UpperCAmelCase ( self : Optional[int] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" _UpperCamelCase = self._add_eos_if_not_present(lowercase_) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(lowercase_) return token_ids_a + token_ids_a def __getstate__( self : Tuple) -> Any: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : Optional[Any] , lowercase_ : Any) -> Optional[int]: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : int , lowercase_ : "TextInput" , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = SPIECE_UNDERLINE + text.replace(lowercase_ , " ") return super().tokenize(lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : int , **lowercase_ : Optional[int]) -> List[str]: """simple docstring""" if not self.legacy: _UpperCamelCase = text.startswith(lowercase_) if is_first: _UpperCamelCase = text[1:] _UpperCamelCase = self.sp_model.encode(lowercase_ , out_type=lowercase_) if not self.legacy and not is_first and not text.startswith(" ") and tokens[0].startswith(lowercase_): _UpperCamelCase = ([tokens[0][1:]] if len(tokens[0]) > 1 else []) + tokens[1:] return tokens def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if token.startswith("<extra_id_"): _UpperCamelCase = re.match(R"<extra_id_(\d+)>" , lowercase_) _UpperCamelCase = int(match.group(1)) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase_) def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> int: """simple docstring""" if index < self.sp_model.get_piece_size(): _UpperCamelCase = self.sp_model.IdToPiece(lowercase_) else: _UpperCamelCase = f'<extra_id_{self.vocab_size - 1 - index}>' return token def __UpperCAmelCase ( self : Dict , lowercase_ : Optional[int]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,)
82
0
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
712
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return getitem, k def lowerCAmelCase__ ( a__ , a__ ) ->Tuple: '''simple docstring''' return setitem, k, v def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' return delitem, k def lowerCAmelCase__ ( a__ , a__ , *a__ ) ->List[str]: '''simple docstring''' try: return fun(a__ , *a__ ), None except Exception as e: return None, e lowerCamelCase__ = ( _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), ) lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_a''', '''val_b'''), ] lowerCamelCase__ = [ _set('''key_a''', '''val_a'''), _set('''key_b''', '''val_b'''), _del('''key_a'''), _del('''key_b'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), ] lowerCamelCase__ = [ _get('''key_a'''), _del('''key_a'''), _set('''key_a''', '''val_a'''), _del('''key_a'''), _del('''key_a'''), _get('''key_a'''), ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize ] lowerCamelCase__ = [ *[_set(x, x) for x in range(5)], # guaranteed upsize *[_del(x) for x in range(5)], _set('''key_a''', '''val_b'''), ] @pytest.mark.parametrize( "operations" , ( pytest.param(_add_items , id="add items" ), pytest.param(_overwrite_items , id="overwrite items" ), pytest.param(_delete_items , id="delete items" ), pytest.param(_access_absent_items , id="access absent items" ), pytest.param(_add_with_resize_up , id="add with resize up" ), pytest.param(_add_with_resize_down , id="add with resize down" ), ) , ) def lowerCAmelCase__ ( a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = HashMap(initial_block_size=4 ) _UpperCamelCase = {} for _, (fun, *args) in enumerate(a__ ): _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) _UpperCamelCase , _UpperCamelCase = _run_operation(a__ , a__ , *a__ ) assert my_res == py_res assert str(a__ ) == str(a__ ) assert set(a__ ) == set(a__ ) assert len(a__ ) == len(a__ ) assert set(my.items() ) == set(py.items() ) def lowerCAmelCase__ ( ) ->List[Any]: '''simple docstring''' def is_public(a__ ) -> bool: return not name.startswith("_" ) _UpperCamelCase = {name for name in dir({} ) if is_public(a__ )} _UpperCamelCase = {name for name in dir(HashMap() ) if is_public(a__ )} assert dict_public_names > hash_public_names
82
0
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str , lowercase_ : int=13 , lowercase_ : str=30 , lowercase_ : Tuple=2 , lowercase_ : Tuple=3 , lowercase_ : Tuple=True , lowercase_ : Tuple=True , lowercase_ : Union[str, Any]=32 , lowercase_ : str=2 , lowercase_ : Optional[int]=4 , lowercase_ : Dict=37 , lowercase_ : Optional[int]="gelu" , lowercase_ : Dict=0.1 , lowercase_ : List[str]=0.1 , lowercase_ : Tuple=10 , lowercase_ : Dict=0.02 , lowercase_ : Optional[Any]=3 , lowercase_ : str=None , ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCamelCase = (image_size // patch_size) ** 2 _UpperCamelCase = num_patches + 1 def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : Optional[Any]) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : Any) -> Tuple: """simple docstring""" _UpperCamelCase = TFViTModel(config=lowercase_) _UpperCamelCase = model(lowercase_ , training=lowercase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) # Test with an image with different size than the one specified in config. _UpperCamelCase = self.image_size // 2 _UpperCamelCase = pixel_values[:, :, :image_size, :image_size] _UpperCamelCase = model(lowercase_ , interpolate_pos_encoding=lowercase_ , training=lowercase_) _UpperCamelCase = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size)) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : Any , lowercase_ : Dict) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.type_sequence_label_size _UpperCamelCase = TFViTForImageClassification(lowercase_) _UpperCamelCase = model(lowercase_ , labels=lowercase_ , training=lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) # Test with an image with different size than the one specified in config. _UpperCamelCase = self.image_size // 2 _UpperCamelCase = pixel_values[:, :, :image_size, :image_size] _UpperCamelCase = model(lowercase_ , interpolate_pos_encoding=lowercase_ , training=lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) # test greyscale images _UpperCamelCase = 1 _UpperCamelCase = TFViTForImageClassification(lowercase_) _UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) _UpperCamelCase = model(lowercase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = config_and_inputs _UpperCamelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( lowerCAmelCase, lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __A = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) __A = False __A = False __A = False def __UpperCAmelCase ( self : Union[str, Any]) -> List[str]: """simple docstring""" _UpperCamelCase = TFViTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37) def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds") def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" pass @unittest.skip(reason="ViT does not use inputs_embeds") def __UpperCAmelCase ( self : Union[str, Any]) -> Tuple: """simple docstring""" pass def __UpperCAmelCase ( self : str) -> List[str]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , tf.keras.layers.Layer)) def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(lowercase_) _UpperCamelCase = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_) def __UpperCAmelCase ( self : Tuple) -> Optional[int]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_) @slow def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = TFViTModel.from_pretrained("google/vit-base-patch16-224") self.assertIsNotNone(lowercase_) def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __UpperCAmelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224") if is_vision_available() else None @slow def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = TFViTForImageClassification.from_pretrained("google/vit-base-patch16-224") _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=lowercase_ , return_tensors="tf") # forward pass _UpperCamelCase = model(**lowercase_) # verify the logits _UpperCamelCase = tf.TensorShape((1, 1000)) self.assertEqual(outputs.logits.shape , lowercase_) _UpperCamelCase = tf.constant([-0.27_44, 0.82_15, -0.08_36]) tf.debugging.assert_near(outputs.logits[0, :3] , lowercase_ , atol=1e-4)
713
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = KandinskyVaaControlnetImgaImgPipeline __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] __A = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] __A = False @property def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" return 32 @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.time_input_dim @property def __UpperCAmelCase ( self : List[str]) -> Any: """simple docstring""" return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : Optional[Any]) -> Union[str, Any]: """simple docstring""" return 100 @property def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCamelCase = UNetaDConditionModel(**lowercase_) return model @property def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : int) -> Dict: """simple docstring""" torch.manual_seed(0) _UpperCamelCase = VQModel(**self.dummy_movq_kwargs) return model def __UpperCAmelCase ( self : int) -> Any: """simple docstring""" _UpperCamelCase = self.dummy_unet _UpperCamelCase = self.dummy_movq _UpperCamelCase = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_00_85, "beta_end": 0.0_12, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _UpperCamelCase = DDIMScheduler(**lowercase_) _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCAmelCase ( self : str , lowercase_ : Dict , lowercase_ : List[str]=0) -> List[str]: """simple docstring""" _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( lowercase_) # create init_image _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] _UpperCamelCase = Image.fromarray(np.uinta(lowercase_)).convert("RGB").resize((256, 256)) # create hint _UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_)).to(lowercase_) if str(lowercase_).startswith("mps"): _UpperCamelCase = torch.manual_seed(lowercase_) else: _UpperCamelCase = torch.Generator(device=lowercase_).manual_seed(lowercase_) _UpperCamelCase = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def __UpperCAmelCase ( self : Any) -> str: """simple docstring""" _UpperCamelCase = "cpu" _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**lowercase_) _UpperCamelCase = pipe.to(lowercase_) pipe.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = pipe(**self.get_dummy_inputs(lowercase_)) _UpperCamelCase = output.images _UpperCamelCase = pipe( **self.get_dummy_inputs(lowercase_) , return_dict=lowercase_ , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Union[str, Any]) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Optional[int]) -> Any: """simple docstring""" _UpperCamelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _UpperCamelCase = init_image.resize((512, 512)) _UpperCamelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _UpperCamelCase = torch.from_numpy(np.array(lowercase_)).float() / 2_55.0 _UpperCamelCase = hint.permute(2 , 0 , 1).unsqueeze(0) _UpperCamelCase = "A robot, 4k photo" _UpperCamelCase = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(lowercase_) _UpperCamelCase = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(lowercase_) pipeline.set_progress_bar_config(disable=lowercase_) _UpperCamelCase = torch.Generator(device="cpu").manual_seed(0) _UpperCamelCase , _UpperCamelCase = pipe_prior( lowercase_ , image=lowercase_ , strength=0.85 , generator=lowercase_ , negative_prompt="" , ).to_tuple() _UpperCamelCase = pipeline( image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , hint=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) _UpperCamelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowercase_ , lowercase_)
82
0
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = 42 __A = jnp.floataa __A = True def __UpperCAmelCase ( self : Any) -> Union[str, Any]: """simple docstring""" super().setup() _UpperCamelCase = nn.Dense(5 , dtype=self.dtype) def __call__( self : Dict , *lowercase_ : Optional[Any] , **lowercase_ : Dict) -> str: """simple docstring""" _UpperCamelCase = super().__call__(*lowercase_ , **lowercase_) _UpperCamelCase = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = FlaxBigBirdForNaturalQuestionsModule def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ , a__ ) ->List[Any]: '''simple docstring''' def cross_entropy(a__ , a__ , a__=None ): _UpperCamelCase = logits.shape[-1] _UpperCamelCase = (labels[..., None] == jnp.arange(a__ )[None]).astype("f4" ) _UpperCamelCase = jax.nn.log_softmax(a__ , axis=-1 ) _UpperCamelCase = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: _UpperCamelCase = reduction(a__ ) return loss _UpperCamelCase = partial(a__ , reduction=jnp.mean ) _UpperCamelCase = cross_entropy(a__ , a__ ) _UpperCamelCase = cross_entropy(a__ , a__ ) _UpperCamelCase = cross_entropy(a__ , a__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class _UpperCAmelCase : '''simple docstring''' __A = '''google/bigbird-roberta-base''' __A = 3_000 __A = 10_500 __A = 128 __A = 3 __A = 1 __A = 5 # tx_args __A = 3e-5 __A = 0.0 __A = 20_000 __A = 0.0_095 __A = '''bigbird-roberta-natural-questions''' __A = '''training-expt''' __A = '''data/nq-training.jsonl''' __A = '''data/nq-validation.jsonl''' def __UpperCAmelCase ( self : str) -> Optional[Any]: """simple docstring""" os.makedirs(self.base_dir , exist_ok=lowercase_) _UpperCamelCase = os.path.join(self.base_dir , self.save_dir) _UpperCamelCase = self.batch_size_per_device * jax.device_count() @dataclass class _UpperCAmelCase : '''simple docstring''' __A = 42 __A = 4_096 # no dynamic padding on TPUs def __call__( self : List[str] , lowercase_ : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = self.collate_fn(lowercase_) _UpperCamelCase = jax.tree_util.tree_map(lowercase_ , lowercase_) return batch def __UpperCAmelCase ( self : str , lowercase_ : str) -> List[str]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.fetch_inputs(features["input_ids"]) _UpperCamelCase = { "input_ids": jnp.array(lowercase_ , dtype=jnp.intaa), "attention_mask": jnp.array(lowercase_ , dtype=jnp.intaa), "start_labels": jnp.array(features["start_token"] , dtype=jnp.intaa), "end_labels": jnp.array(features["end_token"] , dtype=jnp.intaa), "pooled_labels": jnp.array(features["category"] , dtype=jnp.intaa), } return batch def __UpperCAmelCase ( self : List[Any] , lowercase_ : list) -> Optional[int]: """simple docstring""" _UpperCamelCase = [self._fetch_inputs(lowercase_) for ids in input_ids] return zip(*lowercase_) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : list) -> List[str]: """simple docstring""" _UpperCamelCase = [1 for _ in range(len(lowercase_))] while len(lowercase_) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def lowerCAmelCase__ ( a__ , a__ , a__=None ) ->Optional[int]: '''simple docstring''' if seed is not None: _UpperCamelCase = dataset.shuffle(seed=a__ ) for i in range(len(a__ ) // batch_size ): _UpperCamelCase = dataset[i * batch_size : (i + 1) * batch_size] yield dict(a__ ) @partial(jax.pmap , axis_name="batch" ) def lowerCAmelCase__ ( a__ , a__ , **a__ ) ->Any: '''simple docstring''' def loss_fn(a__ ): _UpperCamelCase = model_inputs.pop("start_labels" ) _UpperCamelCase = model_inputs.pop("end_labels" ) _UpperCamelCase = model_inputs.pop("pooled_labels" ) _UpperCamelCase = state.apply_fn(**a__ , params=a__ , dropout_rng=a__ , train=a__ ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = outputs return state.loss_fn( a__ , a__ , a__ , a__ , a__ , a__ , ) _UpperCamelCase , _UpperCamelCase = jax.random.split(a__ ) _UpperCamelCase = jax.value_and_grad(a__ ) _UpperCamelCase , _UpperCamelCase = grad_fn(state.params ) _UpperCamelCase = jax.lax.pmean({"loss": loss} , axis_name="batch" ) _UpperCamelCase = jax.lax.pmean(a__ , "batch" ) _UpperCamelCase = state.apply_gradients(grads=a__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="batch" ) def lowerCAmelCase__ ( a__ , **a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = model_inputs.pop("start_labels" ) _UpperCamelCase = model_inputs.pop("end_labels" ) _UpperCamelCase = model_inputs.pop("pooled_labels" ) _UpperCamelCase = state.apply_fn(**a__ , params=state.params , train=a__ ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = outputs _UpperCamelCase = state.loss_fn(a__ , a__ , a__ , a__ , a__ , a__ ) _UpperCamelCase = jax.lax.pmean({"loss": loss} , axis_name="batch" ) return metrics class _UpperCAmelCase ( train_state.TrainState ): '''simple docstring''' __A = struct.field(pytree_node=lowerCAmelCase ) @dataclass class _UpperCAmelCase : '''simple docstring''' __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = None def __UpperCAmelCase ( self : Tuple , lowercase_ : Any , lowercase_ : List[str] , lowercase_ : int , lowercase_ : Tuple=None) -> Dict: """simple docstring""" _UpperCamelCase = model.params _UpperCamelCase = TrainState.create( apply_fn=model.__call__ , params=lowercase_ , tx=lowercase_ , loss_fn=lowercase_ , ) if ckpt_dir is not None: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = restore_checkpoint(lowercase_ , lowercase_) _UpperCamelCase = { "lr": args.lr, "init_lr": args.init_lr, "warmup_steps": args.warmup_steps, "num_train_steps": num_train_steps, "weight_decay": args.weight_decay, } _UpperCamelCase , _UpperCamelCase = build_tx(**lowercase_) _UpperCamelCase = train_state.TrainState( step=lowercase_ , apply_fn=model.__call__ , params=lowercase_ , tx=lowercase_ , opt_state=lowercase_ , ) _UpperCamelCase = args _UpperCamelCase = data_collator _UpperCamelCase = lr _UpperCamelCase = params _UpperCamelCase = jax_utils.replicate(lowercase_) return state def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple , lowercase_ : List[str] , lowercase_ : Union[str, Any]) -> Any: """simple docstring""" _UpperCamelCase = self.args _UpperCamelCase = len(lowercase_) // args.batch_size _UpperCamelCase = jax.random.PRNGKey(0) _UpperCamelCase = jax.random.split(lowercase_ , jax.device_count()) for epoch in range(args.max_epochs): _UpperCamelCase = jnp.array(0 , dtype=jnp.floataa) _UpperCamelCase = get_batched_dataset(lowercase_ , args.batch_size , seed=lowercase_) _UpperCamelCase = 0 for batch in tqdm(lowercase_ , total=lowercase_ , desc=f'Running EPOCH-{epoch}'): _UpperCamelCase = self.data_collator(lowercase_) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.train_step_fn(lowercase_ , lowercase_ , **lowercase_) running_loss += jax_utils.unreplicate(metrics["loss"]) i += 1 if i % args.logging_steps == 0: _UpperCamelCase = jax_utils.unreplicate(state.step) _UpperCamelCase = running_loss.item() / i _UpperCamelCase = self.scheduler_fn(state_step - 1) _UpperCamelCase = self.evaluate(lowercase_ , lowercase_) _UpperCamelCase = { "step": state_step.item(), "eval_loss": eval_loss.item(), "tr_loss": tr_loss, "lr": lr.item(), } tqdm.write(str(lowercase_)) self.logger.log(lowercase_ , commit=lowercase_) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase_) def __UpperCAmelCase ( self : Tuple , lowercase_ : Optional[int] , lowercase_ : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = get_batched_dataset(lowercase_ , self.args.batch_size) _UpperCamelCase = len(lowercase_) // self.args.batch_size _UpperCamelCase = jnp.array(0 , dtype=jnp.floataa) _UpperCamelCase = 0 for batch in tqdm(lowercase_ , total=lowercase_ , desc="Evaluating ... "): _UpperCamelCase = self.data_collator(lowercase_) _UpperCamelCase = self.val_step_fn(lowercase_ , **lowercase_) running_loss += jax_utils.unreplicate(metrics["loss"]) i += 1 return running_loss / i def __UpperCAmelCase ( self : List[str] , lowercase_ : Union[str, Any] , lowercase_ : int) -> List[Any]: """simple docstring""" _UpperCamelCase = jax_utils.unreplicate(lowercase_) print(f'SAVING CHECKPOINT IN {save_dir}' , end=" ... ") self.model_save_fn(lowercase_ , params=state.params) with open(os.path.join(lowercase_ , "opt_state.msgpack") , "wb") as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase_ , "args.joblib")) joblib.dump(self.data_collator , os.path.join(lowercase_ , "data_collator.joblib")) with open(os.path.join(lowercase_ , "training_state.json") , "w") as f: json.dump({"step": state.step.item()} , lowercase_) print("DONE") def lowerCAmelCase__ ( a__ , a__ ) ->List[str]: '''simple docstring''' print(f'RESTORING CHECKPOINT FROM {save_dir}' , end=" ... " ) with open(os.path.join(a__ , "flax_model.msgpack" ) , "rb" ) as f: _UpperCamelCase = from_bytes(state.params , f.read() ) with open(os.path.join(a__ , "opt_state.msgpack" ) , "rb" ) as f: _UpperCamelCase = from_bytes(state.opt_state , f.read() ) _UpperCamelCase = joblib.load(os.path.join(a__ , "args.joblib" ) ) _UpperCamelCase = joblib.load(os.path.join(a__ , "data_collator.joblib" ) ) with open(os.path.join(a__ , "training_state.json" ) , "r" ) as f: _UpperCamelCase = json.load(a__ ) _UpperCamelCase = training_state["step"] print("DONE" ) return params, opt_state, step, args, data_collator def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->str: '''simple docstring''' _UpperCamelCase = num_train_steps - warmup_steps _UpperCamelCase = optax.linear_schedule(init_value=a__ , end_value=a__ , transition_steps=a__ ) _UpperCamelCase = optax.linear_schedule(init_value=a__ , end_value=1e-7 , transition_steps=a__ ) _UpperCamelCase = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ ) ->Any: '''simple docstring''' def weight_decay_mask(a__ ): _UpperCamelCase = traverse_util.flatten_dict(a__ ) _UpperCamelCase = {k: (v[-1] != "bias" and v[-2:] != ("LayerNorm", "scale")) for k, v in params.items()} return traverse_util.unflatten_dict(a__ ) _UpperCamelCase = scheduler_fn(a__ , a__ , a__ , a__ ) _UpperCamelCase = optax.adamw(learning_rate=a__ , weight_decay=a__ , mask=a__ ) return tx, lr
714
def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _UpperCamelCase , _UpperCamelCase = 1, 1 for _ in range(number_of_steps - 1 ): _UpperCamelCase , _UpperCamelCase = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
82
0
from collections import deque class _UpperCAmelCase : '''simple docstring''' def __init__( self : int , lowercase_ : str , lowercase_ : int , lowercase_ : int) -> None: """simple docstring""" _UpperCamelCase = process_name # process name _UpperCamelCase = arrival_time # arrival time of the process # completion time of finished process or last interrupted time _UpperCamelCase = arrival_time _UpperCamelCase = burst_time # remaining burst time _UpperCamelCase = 0 # total time of the process wait in ready queue _UpperCamelCase = 0 # time from arrival time to completion time class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : int , lowercase_ : list[int] , lowercase_ : deque[Process] , lowercase_ : int , ) -> None: """simple docstring""" _UpperCamelCase = number_of_queues # time slice of queues that round robin algorithm applied _UpperCamelCase = time_slices # unfinished process is in this ready_queue _UpperCamelCase = queue # current time _UpperCamelCase = current_time # finished process is in this sequence queue _UpperCamelCase = deque() def __UpperCAmelCase ( self : Optional[int]) -> list[str]: """simple docstring""" _UpperCamelCase = [] for i in range(len(self.finish_queue)): sequence.append(self.finish_queue[i].process_name) return sequence def __UpperCAmelCase ( self : Dict , lowercase_ : list[Process]) -> list[int]: """simple docstring""" _UpperCamelCase = [] for i in range(len(lowercase_)): waiting_times.append(queue[i].waiting_time) return waiting_times def __UpperCAmelCase ( self : Dict , lowercase_ : list[Process]) -> list[int]: """simple docstring""" _UpperCamelCase = [] for i in range(len(lowercase_)): turnaround_times.append(queue[i].turnaround_time) return turnaround_times def __UpperCAmelCase ( self : str , lowercase_ : list[Process]) -> list[int]: """simple docstring""" _UpperCamelCase = [] for i in range(len(lowercase_)): completion_times.append(queue[i].stop_time) return completion_times def __UpperCAmelCase ( self : Any , lowercase_ : deque[Process]) -> list[int]: """simple docstring""" return [q.burst_time for q in queue] def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Process) -> int: """simple docstring""" process.waiting_time += self.current_time - process.stop_time return process.waiting_time def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : deque[Process]) -> deque[Process]: """simple docstring""" _UpperCamelCase = deque() # sequence deque of finished process while len(lowercase_) != 0: _UpperCamelCase = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(lowercase_) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 _UpperCamelCase = 0 # set the process's turnaround time because it is finished _UpperCamelCase = self.current_time - cp.arrival_time # set the completion time _UpperCamelCase = self.current_time # add the process to queue that has finished queue finished.append(lowercase_) self.finish_queue.extend(lowercase_) # add finished process to finish queue # FCFS will finish all remaining processes return finished def __UpperCAmelCase ( self : List[Any] , lowercase_ : deque[Process] , lowercase_ : int) -> tuple[deque[Process], deque[Process]]: """simple docstring""" _UpperCamelCase = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(lowercase_)): _UpperCamelCase = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(lowercase_) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time _UpperCamelCase = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(lowercase_) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished _UpperCamelCase = 0 # set the finish time _UpperCamelCase = self.current_time # update the process' turnaround time because it is finished _UpperCamelCase = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(lowercase_) self.finish_queue.extend(lowercase_) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def __UpperCAmelCase ( self : Any) -> deque[Process]: """simple docstring""" for i in range(self.number_of_queues - 1): _UpperCamelCase , _UpperCamelCase = self.round_robin( self.ready_queue , self.time_slices[i]) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue) return self.finish_queue if __name__ == "__main__": import doctest lowerCamelCase__ = Process('''P1''', 0, 53) lowerCamelCase__ = Process('''P2''', 0, 17) lowerCamelCase__ = Process('''P3''', 0, 68) lowerCamelCase__ = Process('''P4''', 0, 24) lowerCamelCase__ = 3 lowerCamelCase__ = [17, 25] lowerCamelCase__ = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={'''queue''': deque([Pa, Pa, Pa, Pa])}) lowerCamelCase__ = Process('''P1''', 0, 53) lowerCamelCase__ = Process('''P2''', 0, 17) lowerCamelCase__ = Process('''P3''', 0, 68) lowerCamelCase__ = Process('''P4''', 0, 24) lowerCamelCase__ = 3 lowerCamelCase__ = [17, 25] lowerCamelCase__ = deque([Pa, Pa, Pa, Pa]) lowerCamelCase__ = MLFQ(number_of_queues, time_slices, queue, 0) lowerCamelCase__ = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F"waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}" ) # print completion times of processes(P1, P2, P3, P4) print( F"completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}" ) # print total turnaround times of processes(P1, P2, P3, P4) print( F"turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}" ) # print sequence of finished processes print( F"sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}" )
715
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , **lowercase_ : Tuple) -> Any: """simple docstring""" super().__init__(**lowercase_) if self.framework == "tf": raise ValueError(f'The {self.__class__} is only available in PyTorch.') requires_backends(self , "vision") self.check_model_type(lowercase_) def __call__( self : str , lowercase_ : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase_ : Union[str, List[str]] = None , **lowercase_ : str , ) -> List[str]: """simple docstring""" if "text_queries" in kwargs: _UpperCamelCase = kwargs.pop("text_queries") if isinstance(lowercase_ , (str, Image.Image)): _UpperCamelCase = {"image": image, "candidate_labels": candidate_labels} else: _UpperCamelCase = image _UpperCamelCase = super().__call__(lowercase_ , **lowercase_) return results def __UpperCAmelCase ( self : Any , **lowercase_ : int) -> List[str]: """simple docstring""" _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs["threshold"] if "top_k" in kwargs: _UpperCamelCase = kwargs["top_k"] return {}, {}, postprocess_params def __UpperCAmelCase ( self : List[Any] , lowercase_ : Any) -> List[str]: """simple docstring""" _UpperCamelCase = load_image(inputs["image"]) _UpperCamelCase = inputs["candidate_labels"] if isinstance(lowercase_ , lowercase_): _UpperCamelCase = candidate_labels.split(",") _UpperCamelCase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa) for i, candidate_label in enumerate(lowercase_): _UpperCamelCase = self.tokenizer(lowercase_ , return_tensors=self.framework) _UpperCamelCase = self.image_processor(lowercase_ , return_tensors=self.framework) yield { "is_last": i == len(lowercase_) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __UpperCAmelCase ( self : Dict , lowercase_ : Tuple) -> str: """simple docstring""" _UpperCamelCase = model_inputs.pop("target_size") _UpperCamelCase = model_inputs.pop("candidate_label") _UpperCamelCase = model_inputs.pop("is_last") _UpperCamelCase = self.model(**lowercase_) _UpperCamelCase = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __UpperCAmelCase ( self : int , lowercase_ : Tuple , lowercase_ : List[str]=0.1 , lowercase_ : int=None) -> List[str]: """simple docstring""" _UpperCamelCase = [] for model_output in model_outputs: _UpperCamelCase = model_output["candidate_label"] _UpperCamelCase = BaseModelOutput(lowercase_) _UpperCamelCase = self.image_processor.post_process_object_detection( outputs=lowercase_ , threshold=lowercase_ , target_sizes=model_output["target_size"])[0] for index in outputs["scores"].nonzero(): _UpperCamelCase = outputs["scores"][index].item() _UpperCamelCase = self._get_bounding_box(outputs["boxes"][index][0]) _UpperCamelCase = {"score": score, "label": label, "box": box} results.append(lowercase_) _UpperCamelCase = sorted(lowercase_ , key=lambda lowercase_: x["score"] , reverse=lowercase_) if top_k: _UpperCamelCase = results[:top_k] return results def __UpperCAmelCase ( self : str , lowercase_ : "torch.Tensor") -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch.") _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
0
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
716
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __UpperCAmelCase ( self : List[str] , lowercase_ : str) -> str: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") _UpperCamelCase = input_file.read() _UpperCamelCase = regexp.search(lowercase_) return match def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> int: """simple docstring""" with open(lowercase_ , encoding="utf-8") as input_file: _UpperCamelCase = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) _UpperCamelCase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _UpperCamelCase = regexp.finditer(lowercase_) _UpperCamelCase = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def __UpperCAmelCase ( self : int) -> int: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase_)): raise AssertionError(f'open(...) must use utf-8 encoding in {dataset}') def __UpperCAmelCase ( self : str) -> str: """simple docstring""" _UpperCamelCase = Path("./datasets") _UpperCamelCase = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(lowercase_)): raise AssertionError(f'print statement found in {dataset}. Use datasets.logger/logging instead.')
82
0
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def lowerCAmelCase__ ( a__ , a__ ) ->np.array: '''simple docstring''' _UpperCamelCase = f'{sampling_rate}' _UpperCamelCase = "1" _UpperCamelCase = "f32le" _UpperCamelCase = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(a__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _UpperCamelCase = ffmpeg_process.communicate(a__ ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error _UpperCamelCase = output_stream[0] _UpperCamelCase = np.frombuffer(a__ , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def lowerCAmelCase__ ( a__ , a__ , a__ = "f32le" , ) ->int: '''simple docstring''' _UpperCamelCase = f'{sampling_rate}' _UpperCamelCase = "1" if format_for_conversion == "s16le": _UpperCamelCase = 2 elif format_for_conversion == "f32le": _UpperCamelCase = 4 else: raise ValueError(f'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) _UpperCamelCase = platform.system() if system == "Linux": _UpperCamelCase = "alsa" _UpperCamelCase = "default" elif system == "Darwin": _UpperCamelCase = "avfoundation" _UpperCamelCase = ":0" elif system == "Windows": _UpperCamelCase = "dshow" _UpperCamelCase = "default" _UpperCamelCase = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] _UpperCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _UpperCamelCase = _ffmpeg_stream(a__ , a__ ) for item in iterator: yield item def lowerCAmelCase__ ( a__ , a__ , a__ = None , a__ = None , a__ = "f32le" , ) ->Union[str, Any]: '''simple docstring''' if stream_chunk_s is not None: _UpperCamelCase = stream_chunk_s else: _UpperCamelCase = chunk_length_s _UpperCamelCase = ffmpeg_microphone(a__ , a__ , format_for_conversion=a__ ) if format_for_conversion == "s16le": _UpperCamelCase = np.intaa _UpperCamelCase = 2 elif format_for_conversion == "f32le": _UpperCamelCase = np.floataa _UpperCamelCase = 4 else: raise ValueError(f'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) if stride_length_s is None: _UpperCamelCase = chunk_length_s / 6 _UpperCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(a__ , (int, float) ): _UpperCamelCase = [stride_length_s, stride_length_s] _UpperCamelCase = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _UpperCamelCase = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _UpperCamelCase = datetime.datetime.now() _UpperCamelCase = datetime.timedelta(seconds=a__ ) for item in chunk_bytes_iter(a__ , a__ , stride=(stride_left, stride_right) , stream=a__ ): # Put everything back in numpy scale _UpperCamelCase = np.frombuffer(item["raw"] , dtype=a__ ) _UpperCamelCase = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) _UpperCamelCase = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def lowerCAmelCase__ ( a__ , a__ , a__ , a__ = False ) ->Union[str, Any]: '''simple docstring''' _UpperCamelCase = B"" _UpperCamelCase , _UpperCamelCase = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}' ) _UpperCamelCase = 0 for raw in iterator: acc += raw if stream and len(a__ ) < chunk_len: _UpperCamelCase = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(a__ ) >= chunk_len: # We are flushing the accumulator _UpperCamelCase = (_stride_left, stride_right) _UpperCamelCase = {"raw": acc[:chunk_len], "stride": stride} if stream: _UpperCamelCase = False yield item _UpperCamelCase = stride_left _UpperCamelCase = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(a__ ) > stride_left: _UpperCamelCase = {"raw": acc, "stride": (_stride_left, 0)} if stream: _UpperCamelCase = False yield item def lowerCAmelCase__ ( a__ , a__ ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = 2**24 # 16Mo try: with subprocess.Popen(a__ , stdout=subprocess.PIPE , bufsize=a__ ) as ffmpeg_process: while True: _UpperCamelCase = ffmpeg_process.stdout.read(a__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , lowercase_ : str = None , lowercase_ : uuid.UUID = None , lowercase_ : List[Any]=None , lowercase_ : int=None) -> Dict: """simple docstring""" if not conversation_id: _UpperCamelCase = uuid.uuida() if past_user_inputs is None: _UpperCamelCase = [] if generated_responses is None: _UpperCamelCase = [] _UpperCamelCase = conversation_id _UpperCamelCase = past_user_inputs _UpperCamelCase = generated_responses _UpperCamelCase = text def __eq__( self : Optional[Any] , lowercase_ : Optional[Any]) -> List[Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : bool = False) -> Any: """simple docstring""" if self.new_user_input: if overwrite: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' f'with: "{text}".') _UpperCamelCase = text else: logger.warning( f'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' f'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input') else: _UpperCamelCase = text def __UpperCAmelCase ( self : Optional[int]) -> List[Any]: """simple docstring""" if self.new_user_input: self.past_user_inputs.append(self.new_user_input) _UpperCamelCase = None def __UpperCAmelCase ( self : Dict , lowercase_ : str) -> Optional[Any]: """simple docstring""" self.generated_responses.append(lowercase_) def __UpperCAmelCase ( self : List[Any]) -> Optional[int]: """simple docstring""" for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any]) -> int: """simple docstring""" _UpperCamelCase = f'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): _UpperCamelCase = "user" if is_user else "bot" output += f'{name} >> {text} \n' return output @add_end_docstrings( lowerCAmelCase, R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''', ) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[Any] , *lowercase_ : Optional[Any] , **lowercase_ : str) -> List[str]: """simple docstring""" super().__init__(*lowercase_ , **lowercase_) if self.tokenizer.pad_token_id is None: _UpperCamelCase = self.tokenizer.eos_token def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Union[str, Any]=None , lowercase_ : int=None , lowercase_ : str=None , **lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} if min_length_for_response is not None: _UpperCamelCase = min_length_for_response if minimum_tokens is not None: _UpperCamelCase = minimum_tokens if "max_length" in generate_kwargs: _UpperCamelCase = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _UpperCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowercase_) return preprocess_params, forward_params, postprocess_params def __call__( self : Any , lowercase_ : Union[Conversation, List[Conversation]] , lowercase_ : str=0 , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = super().__call__(lowercase_ , num_workers=lowercase_ , **lowercase_) if isinstance(lowercase_ , lowercase_) and len(lowercase_) == 1: return outputs[0] return outputs def __UpperCAmelCase ( self : List[Any] , lowercase_ : Conversation , lowercase_ : Any=32) -> Dict[str, Any]: """simple docstring""" if not isinstance(lowercase_ , lowercase_): raise ValueError("ConversationalPipeline, expects Conversation as inputs") if conversation.new_user_input is None: raise ValueError( f'Conversation with UUID {type(conversation.uuid)} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method") if hasattr(self.tokenizer , "_build_conversation_input_ids"): _UpperCamelCase = self.tokenizer._build_conversation_input_ids(lowercase_) else: # If the tokenizer cannot handle conversations, we default to only the old version _UpperCamelCase = self._legacy_parse_and_tokenize(lowercase_) if self.framework == "pt": _UpperCamelCase = torch.LongTensor([input_ids]) elif self.framework == "tf": _UpperCamelCase = tf.constant([input_ids]) return {"input_ids": input_ids, "conversation": conversation} def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[int]=10 , **lowercase_ : Dict) -> List[str]: """simple docstring""" _UpperCamelCase = generate_kwargs.get("max_length" , self.model.config.max_length) _UpperCamelCase = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})') _UpperCamelCase = max_length - minimum_tokens _UpperCamelCase = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: _UpperCamelCase = model_inputs["attention_mask"][:, -trim:] _UpperCamelCase = model_inputs.pop("conversation") _UpperCamelCase = max_length _UpperCamelCase = self.model.generate(**lowercase_ , **lowercase_) if self.model.config.is_encoder_decoder: _UpperCamelCase = 1 else: _UpperCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : int=True) -> List[Any]: """simple docstring""" _UpperCamelCase = model_outputs["output_ids"] _UpperCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) _UpperCamelCase = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowercase_) return conversation def __UpperCAmelCase ( self : Any , lowercase_ : Conversation) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.eos_token_id _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_) + [eos_token_id]) else: input_ids.extend(self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_)) if len(lowercase_) > self.tokenizer.model_max_length: _UpperCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
82
0
from __future__ import annotations def lowerCAmelCase__ ( a__ , a__ = None , a__ = None , a__ = False , ) ->tuple[int, float, str]: '''simple docstring''' _UpperCamelCase = cipher_alphabet or [chr(a__ ) for i in range(97 , 123 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) _UpperCamelCase = { "a": 0.08497, "b": 0.01492, "c": 0.02202, "d": 0.04253, "e": 0.11162, "f": 0.02228, "g": 0.02015, "h": 0.06094, "i": 0.07546, "j": 0.00153, "k": 0.01292, "l": 0.04025, "m": 0.02406, "n": 0.06749, "o": 0.07507, "p": 0.01929, "q": 0.00095, "r": 0.07587, "s": 0.06327, "t": 0.09356, "u": 0.02758, "v": 0.00978, "w": 0.02560, "x": 0.00150, "y": 0.01994, "z": 0.00077, } else: # Custom frequencies dictionary _UpperCamelCase = frequencies_dict if not case_sensitive: _UpperCamelCase = ciphertext.lower() # Chi squared statistic values _UpperCamelCase = {} # cycle through all of the shifts for shift in range(len(a__ ) ): _UpperCamelCase = "" # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet _UpperCamelCase = (alphabet_letters.index(letter.lower() ) - shift) % len( a__ ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter _UpperCamelCase = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: _UpperCamelCase = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message _UpperCamelCase = decrypted_with_shift.lower().count(a__ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _UpperCamelCase = frequencies[letter] * occurrences # Complete the chi squared statistic formula _UpperCamelCase = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message _UpperCamelCase = decrypted_with_shift.count(a__ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _UpperCamelCase = frequencies[letter] * occurrences # Complete the chi squared statistic formula _UpperCamelCase = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary _UpperCamelCase = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(a__ ) -> tuple[float, str]: return chi_squared_statistic_values[key] _UpperCamelCase = min( a__ , key=a__ , ) # Get all the data from the most likely cipher (key, decoded message) ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
718
def lowerCAmelCase__ ( a__ = 50 ) ->int: '''simple docstring''' _UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
82
0
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def lowerCAmelCase__ ( a__ , a__ ) ->List[Any]: '''simple docstring''' _UpperCamelCase = old_name if "patch_embed" in old_name: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = old_name.split("." ) if layer == "0": _UpperCamelCase = old_name.replace("0" , "convolution1" ) elif layer == "1": _UpperCamelCase = old_name.replace("1" , "batchnorm_before" ) elif layer == "3": _UpperCamelCase = old_name.replace("3" , "convolution2" ) else: _UpperCamelCase = old_name.replace("4" , "batchnorm_after" ) if "network" in old_name and re.search(r"\d\.\d" , a__ ): _UpperCamelCase = r"\b\d{2}\b" if bool(re.search(a__ , a__ ) ): _UpperCamelCase = re.search(r"\d\.\d\d." , a__ ).group() else: _UpperCamelCase = re.search(r"\d\.\d." , a__ ).group() if int(match[0] ) < 6: _UpperCamelCase = old_name.replace(a__ , "" ) _UpperCamelCase = trimmed_name.replace("network" , match[0] + ".meta4D_layers.blocks." + match[2:-1] ) _UpperCamelCase = "intermediate_stages." + trimmed_name else: _UpperCamelCase = old_name.replace(a__ , "" ) if int(match[2] ) < num_meta4D_last_stage: _UpperCamelCase = trimmed_name.replace("network" , "meta4D_layers.blocks." + match[2] ) else: _UpperCamelCase = str(int(match[2] ) - num_meta4D_last_stage ) _UpperCamelCase = trimmed_name.replace("network" , "meta3D_layers.blocks." + layer_index ) if "norm1" in old_name: _UpperCamelCase = trimmed_name.replace("norm1" , "layernorm1" ) elif "norm2" in old_name: _UpperCamelCase = trimmed_name.replace("norm2" , "layernorm2" ) elif "fc1" in old_name: _UpperCamelCase = trimmed_name.replace("fc1" , "linear_in" ) elif "fc2" in old_name: _UpperCamelCase = trimmed_name.replace("fc2" , "linear_out" ) _UpperCamelCase = "last_stage." + trimmed_name elif "network" in old_name and re.search(r".\d." , a__ ): _UpperCamelCase = old_name.replace("network" , "intermediate_stages" ) if "fc" in new_name: _UpperCamelCase = new_name.replace("fc" , "convolution" ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): _UpperCamelCase = new_name.replace("norm1" , "batchnorm_before" ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): _UpperCamelCase = new_name.replace("norm2" , "batchnorm_after" ) if "proj" in new_name: _UpperCamelCase = new_name.replace("proj" , "projection" ) if "dist_head" in new_name: _UpperCamelCase = new_name.replace("dist_head" , "distillation_classifier" ) elif "head" in new_name: _UpperCamelCase = new_name.replace("head" , "classifier" ) elif "patch_embed" in new_name: _UpperCamelCase = "efficientformer." + new_name elif new_name == "norm.weight" or new_name == "norm.bias": _UpperCamelCase = new_name.replace("norm" , "layernorm" ) _UpperCamelCase = "efficientformer." + new_name else: _UpperCamelCase = "efficientformer.encoder." + new_name return new_name def lowerCAmelCase__ ( a__ , a__ ) ->List[str]: '''simple docstring''' for key in checkpoint.copy().keys(): _UpperCamelCase = checkpoint.pop(a__ ) _UpperCamelCase = val return checkpoint def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCamelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return image def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->List[Any]: '''simple docstring''' _UpperCamelCase = torch.load(a__ , map_location="cpu" )["model"] _UpperCamelCase = EfficientFormerConfig.from_json_file(a__ ) _UpperCamelCase = EfficientFormerForImageClassificationWithTeacher(a__ ) _UpperCamelCase = "_".join(checkpoint_path.split("/" )[-1].split("." )[0].split("_" )[:-1] ) _UpperCamelCase = config.depths[-1] - config.num_metaad_blocks + 1 _UpperCamelCase = convert_torch_checkpoint(a__ , a__ ) model.load_state_dict(a__ ) model.eval() _UpperCamelCase = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } # prepare image _UpperCamelCase = prepare_img() _UpperCamelCase = 256 _UpperCamelCase = 224 _UpperCamelCase = EfficientFormerImageProcessor( size={"shortest_edge": image_size} , crop_size={"height": crop_size, "width": crop_size} , resample=pillow_resamplings["bicubic"] , ) _UpperCamelCase = processor(images=a__ , return_tensors="pt" ).pixel_values # original processing pipeline _UpperCamelCase = Compose( [ Resize(a__ , interpolation=pillow_resamplings["bicubic"] ), CenterCrop(a__ ), ToTensor(), Normalize(a__ , a__ ), ] ) _UpperCamelCase = image_transforms(a__ ).unsqueeze(0 ) assert torch.allclose(a__ , a__ ) _UpperCamelCase = model(a__ ) _UpperCamelCase = outputs.logits _UpperCamelCase = (1, 1_000) if "l1" in model_name: _UpperCamelCase = torch.Tensor( [-0.1312, 0.4353, -1.0499, -0.5124, 0.4183, -0.6793, -1.3777, -0.0893, -0.7358, -2.4328] ) assert torch.allclose(logits[0, :10] , a__ , atol=1e-3 ) assert logits.shape == expected_shape elif "l3" in model_name: _UpperCamelCase = torch.Tensor( [-1.3150, -1.5456, -1.2556, -0.8496, -0.7127, -0.7897, -0.9728, -0.3052, 0.3751, -0.3127] ) assert torch.allclose(logits[0, :10] , a__ , atol=1e-3 ) assert logits.shape == expected_shape elif "l7" in model_name: _UpperCamelCase = torch.Tensor( [-1.0283, -1.4131, -0.5644, -1.3115, -0.5785, -1.2049, -0.7528, 0.1992, -0.3822, -0.0878] ) assert logits.shape == expected_shape else: raise ValueError( f'Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7' ) # Save Checkpoints Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) print(f'Checkpoint successfuly converted. Model saved at {pytorch_dump_path}' ) processor.save_pretrained(a__ ) print(f'Processor successfuly saved at {pytorch_dump_path}' ) if push_to_hub: print("Pushing model to the hub..." ) model.push_to_hub( repo_id=f'Bearnardd/{pytorch_dump_path}' , commit_message="Add model" , use_temp_dir=a__ , ) processor.push_to_hub( repo_id=f'Bearnardd/{pytorch_dump_path}' , commit_message="Add image processor" , use_temp_dir=a__ , ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--pytorch_model_path''', default=None, type=str, required=True, help='''Path to EfficientFormer pytorch checkpoint.''', ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The json file for EfficientFormer model config.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') parser.add_argument( '''--no-push_to_hub''', dest='''push_to_hub''', action='''store_false''', help='''Do not push model and image processor to the hub''', ) parser.set_defaults(push_to_hub=True) lowerCamelCase__ = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
719
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' _UpperCamelCase = 1.5 _UpperCamelCase = int(factor * num_class_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 ) os.makedirs(f'{class_data_dir}/images' , exist_ok=a__ ) if len(list(Path(f'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: _UpperCamelCase = client.query(text=a__ ) if len(a__ ) >= factor * num_class_images or num_images > 1e4: break else: _UpperCamelCase = int(factor * num_images ) _UpperCamelCase = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=a__ , aesthetic_weight=0.1 , ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = tqdm(desc="downloading real regularization images" , total=a__ ) with open(f'{class_data_dir}/caption.txt' , "w" ) as fa, open(f'{class_data_dir}/urls.txt' , "w" ) as fa, open( f'{class_data_dir}/images.txt' , "w" ) as fa: while total < num_class_images: _UpperCamelCase = class_images[count] count += 1 try: _UpperCamelCase = requests.get(images["url"] ) if img.status_code == 200: _UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(f'{class_data_dir}/images/{total}.jpg' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'{class_data_dir}/images/{total}.jpg' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = argparse.ArgumentParser("" , add_help=a__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=a__ , type=a__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=a__ , type=a__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=a__ ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
82
0
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowerCamelCase__ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Dict , lowercase_ : Optional[int] , lowercase_ : str=7 , lowercase_ : str=3 , lowercase_ : List[Any]=18 , lowercase_ : Tuple=30 , lowercase_ : Tuple=400 , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=True , lowercase_ : List[Any]=True , lowercase_ : Dict=None , ) -> List[str]: """simple docstring""" _UpperCamelCase = size if size is not None else {"height": 20, "width": 20} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = size _UpperCamelCase = do_normalize _UpperCamelCase = do_convert_rgb _UpperCamelCase = [512, 1024, 2048, 4096] _UpperCamelCase = patch_size if patch_size is not None else {"height": 16, "width": 16} def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def __UpperCAmelCase ( self : List[Any]) -> str: """simple docstring""" _UpperCamelCase = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" _UpperCamelCase = Image.open(requests.get(lowercase_ , stream=lowercase_).raw).convert("RGB") return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11, reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''', ) @require_torch @require_vision class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = PixaStructImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self : Union[str, Any]) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = PixaStructImageProcessingTester(self) @property def __UpperCAmelCase ( self : Optional[Any]) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self : Optional[int]) -> Tuple: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowercase_ , "do_normalize")) self.assertTrue(hasattr(lowercase_ , "do_convert_rgb")) def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" _UpperCamelCase = self.image_processor_tester.prepare_dummy_image() _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) _UpperCamelCase = 2048 _UpperCamelCase = image_processor(lowercase_ , return_tensors="pt" , max_patches=lowercase_) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.06_06) , atol=1e-3 , rtol=1e-3)) def __UpperCAmelCase ( self : List[Any]) -> Tuple: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , Image.Image) # Test not batched input _UpperCamelCase = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _UpperCamelCase = image_processor( lowercase_ , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __UpperCAmelCase ( self : Any) -> Tuple: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , Image.Image) # Test not batched input _UpperCamelCase = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 _UpperCamelCase = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(lowercase_): _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_).flattened_patches _UpperCamelCase = "Hello" _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_ , header_text=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _UpperCamelCase = image_processor( lowercase_ , return_tensors="pt" , max_patches=lowercase_ , header_text=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __UpperCAmelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , numpify=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , np.ndarray) _UpperCamelCase = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _UpperCamelCase = image_processor( lowercase_ , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , torchify=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , torch.Tensor) # Test not batched input _UpperCamelCase = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _UpperCamelCase = image_processor( lowercase_ , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11, reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''', ) @require_torch @require_vision class _UpperCAmelCase ( lowerCAmelCase, unittest.TestCase ): '''simple docstring''' __A = PixaStructImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self : str) -> Optional[Any]: """simple docstring""" _UpperCamelCase = PixaStructImageProcessingTester(self , num_channels=4) _UpperCamelCase = 3 @property def __UpperCAmelCase ( self : Optional[int]) -> Union[str, Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self : str) -> List[Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowercase_ , "do_normalize")) self.assertTrue(hasattr(lowercase_ , "do_convert_rgb")) def __UpperCAmelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_) for image in image_inputs: self.assertIsInstance(lowercase_ , Image.Image) # Test not batched input _UpperCamelCase = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _UpperCamelCase = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _UpperCamelCase = image_processor( lowercase_ , return_tensors="pt" , max_patches=lowercase_).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
720
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase__ = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' lowerCamelCase__ = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' lowerCamelCase__ = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self : Dict) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string"), "references": datasets.Value("string"), }) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def __UpperCAmelCase ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str) -> Tuple: """simple docstring""" _UpperCamelCase = 0.0 for i, j in zip(lowercase_ , lowercase_): n_correct += 1.0 if math_equivalence.is_equiv(lowercase_ , lowercase_) else 0.0 _UpperCamelCase = n_correct / len(lowercase_) return { "accuracy": accuracy, }
82
0
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , *lowercase_ : Optional[int] , **lowercase_ : List[str]) -> None: """simple docstring""" warnings.warn( "The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use OwlViTImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_)
721
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase__ = 5_0000 lowerCamelCase__ = 5000 lowerCamelCase__,lowerCamelCase__ = os.path.split(__file__) lowerCamelCase__ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase__ ( a__ , a__ ) ->int: '''simple docstring''' for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' for i in range(0 , len(a__ ) , a__ ): _UpperCamelCase = dataset[i : i + batch_size] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(a__ ): _UpperCamelCase = dataset[i] @get_duration def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' with dataset.formatted_as(type=a__ ): for i in range(0 , a__ , a__ ): _UpperCamelCase = dataset[i : i + batch_size] def lowerCAmelCase__ ( ) ->Dict: '''simple docstring''' _UpperCamelCase = {"num examples": SPEED_TEST_N_EXAMPLES} _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] _UpperCamelCase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) _UpperCamelCase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) _UpperCamelCase = generate_example_dataset( os.path.join(a__ , "dataset.arrow" ) , a__ , num_examples=a__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(a__ ) ) _UpperCamelCase = func(a__ , **a__ ) print("shuffling dataset" ) _UpperCamelCase = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(a__ ) ) _UpperCamelCase = func( a__ , **a__ ) with open(a__ , "wb" ) as f: f.write(json.dumps(a__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
0
def snake_case ( snake_case__ :list[int]) -> list[int]: _A = len(snake_case__) for i in range(snake_case__): for j in range(i + 1 , snake_case__): if numbers[j] < numbers[i]: _A , _A = numbers[j], numbers[i] return numbers if __name__ == "__main__": _SCREAMING_SNAKE_CASE = input('Enter numbers separated by a comma:\n').strip() _SCREAMING_SNAKE_CASE = [int(item) for item in user_input.split(',')] print(exchange_sort(unsorted))
83
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _SCREAMING_SNAKE_CASE = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @dataclass class a : """simple docstring""" lowerCamelCase :str = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(glue_processors.keys() )} ) lowerCamelCase :str = field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''} ) lowerCamelCase :int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCamelCase :bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def UpperCAmelCase ( self ) -> str: _A = self.task_name.lower() class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Optional[Any] = '''train''' lowerCamelCase :Optional[int] = '''dev''' lowerCamelCase :Any = '''test''' class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :GlueDataTrainingArguments lowerCamelCase :str lowerCamelCase :List[InputFeatures] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = Split.train , lowerCAmelCase_ = None , ) -> List[Any]: warnings.warn( """This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets """ """library. You can have a look at this example script for pointers: """ """https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py""" , lowerCAmelCase_ , ) _A = args _A = glue_processors[args.task_name]() _A = glue_output_modes[args.task_name] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): try: _A = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) # Load data features from cache or dataset file _A = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}''' , ) _A = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) _A , _A = label_list[2], label_list[1] _A = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _A = cached_features_file + """.lock""" with FileLock(lowerCAmelCase_ ): if os.path.exists(lowerCAmelCase_ ) and not args.overwrite_cache: _A = time.time() _A = torch.load(lowerCAmelCase_ ) logger.info( F'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) else: logger.info(F'''Creating features from dataset file at {args.data_dir}''' ) if mode == Split.dev: _A = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: _A = self.processor.get_test_examples(args.data_dir ) else: _A = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: _A = examples[:limit_length] _A = glue_convert_examples_to_features( lowerCAmelCase_ , lowerCAmelCase_ , max_length=args.max_seq_length , label_list=lowerCAmelCase_ , output_mode=self.output_mode , ) _A = time.time() torch.save(self.features , lowerCAmelCase_ ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self ) -> Tuple: return len(self.features ) def __getitem__( self , lowerCAmelCase_ ) -> InputFeatures: return self.features[i] def UpperCAmelCase ( self ) -> int: return self.label_list
83
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''philschmid/bart-large-cnn-samsum''' lowerCamelCase :Tuple = ( '''This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ''' '''and returns a summary of the text.''' ) lowerCamelCase :List[Any] = '''summarizer''' lowerCamelCase :List[str] = AutoTokenizer lowerCamelCase :Dict = AutoModelForSeqaSeqLM lowerCamelCase :int = ['''text'''] lowerCamelCase :List[Any] = ['''text'''] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: return self.pre_processor(lowerCAmelCase_ , return_tensors="""pt""" , truncation=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: return self.model.generate(**lowerCAmelCase_ )[0] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: return self.pre_processor.decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ )
83
1
import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'artists_file': 'artists.json', 'lyrics_file': 'lyrics.json', 'genres_file': 'genres.json', } _SCREAMING_SNAKE_CASE = { 'artists_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json', }, 'genres_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json', }, 'lyrics_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json', }, } _SCREAMING_SNAKE_CASE = { 'jukebox': 512, } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :int = VOCAB_FILES_NAMES lowerCamelCase :Any = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase :Tuple = PRETRAINED_LYRIC_TOKENS_SIZES lowerCamelCase :Tuple = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=["v3", "v2", "v2"] , lowerCAmelCase_=5_12 , lowerCAmelCase_=5 , lowerCAmelCase_="<|endoftext|>" , **lowerCAmelCase_ , ) -> Tuple: _A = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else unk_token super().__init__( unk_token=lowerCAmelCase_ , n_genres=lowerCAmelCase_ , version=lowerCAmelCase_ , max_n_lyric_tokens=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A = version _A = max_n_lyric_tokens _A = n_genres with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) _A = r"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder ) == 79: _A = oov.replace(r"""\-'""" , r"""\-+'""" ) _A = regex.compile(lowerCAmelCase_ ) _A = {v: k for k, v in self.artists_encoder.items()} _A = {v: k for k, v in self.genres_encoder.items()} _A = {v: k for k, v in self.lyrics_encoder.items()} @property def UpperCAmelCase ( self ) -> Optional[int]: return len(self.artists_encoder ) + len(self.genres_encoder ) + len(self.lyrics_encoder ) def UpperCAmelCase ( self ) -> List[str]: return dict(self.artists_encoder , self.genres_encoder , self.lyrics_encoder ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Dict: _A = [self.artists_encoder.get(lowerCAmelCase_ , 0 ) for artist in list_artists] for genres in range(len(lowerCAmelCase_ ) ): _A = [self.genres_encoder.get(lowerCAmelCase_ , 0 ) for genre in list_genres[genres]] _A = list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres] )) _A = [[self.lyrics_encoder.get(lowerCAmelCase_ , 0 ) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: return list(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> str: _A , _A , _A = self.prepare_for_tokenization(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = self._tokenize(lowerCAmelCase_ ) return artist, genre, lyrics def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = False ) -> Tuple[str, str, str, Dict[str, Any]]: for idx in range(len(self.version ) ): if self.version[idx] == "v3": _A = artists[idx].lower() _A = [genres[idx].lower()] else: _A = self._normalize(artists[idx] ) + """.v2""" _A = [ self._normalize(lowerCAmelCase_ ) + """.v2""" for genre in genres[idx].split("""_""" ) ] # split is for the full dictionary with combined genres if self.version[0] == "v2": _A = regex.compile(r"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" ) _A = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+'\"()[] \t\n""" _A = {vocab[index]: index + 1 for index in range(len(lowerCAmelCase_ ) )} _A = 0 _A = len(lowerCAmelCase_ ) + 1 _A = self.vocab _A = {v: k for k, v in self.vocab.items()} _A = """""" else: _A = regex.compile(r"""[^A-Za-z0-9.,:;!?\-+'\"()\[\] \t\n]+""" ) _A = self._run_strip_accents(lowerCAmelCase_ ) _A = lyrics.replace("""\\""" , """\n""" ) _A = self.out_of_vocab.sub("""""" , lowerCAmelCase_ ), [], [] return artists, genres, lyrics def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = unicodedata.normalize("""NFD""" , lowerCAmelCase_ ) _A = [] for char in text: _A = unicodedata.category(lowerCAmelCase_ ) if cat == "Mn": continue output.append(lowerCAmelCase_ ) return "".join(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = ( [chr(lowerCAmelCase_ ) for i in range(ord("""a""" ) , ord("""z""" ) + 1 )] + [chr(lowerCAmelCase_ ) for i in range(ord("""A""" ) , ord("""Z""" ) + 1 )] + [chr(lowerCAmelCase_ ) for i in range(ord("""0""" ) , ord("""9""" ) + 1 )] + ["""."""] ) _A = frozenset(lowerCAmelCase_ ) _A = re.compile(r"""_+""" ) _A = """""".join([c if c in accepted else """_""" for c in text.lower()] ) _A = pattern.sub("""_""" , lowerCAmelCase_ ).strip("""_""" ) return text def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: return " ".join(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = False ) -> Union[str, Any]: # Convert to TensorType if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = TensorType(lowerCAmelCase_ ) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( """Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.""" ) import tensorflow as tf _A = tf.constant _A = tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError("""Unable to convert output to PyTorch tensors format, PyTorch is not installed.""" ) import torch _A = torch.tensor _A = torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError("""Unable to convert output to JAX tensors format, JAX is not installed.""" ) import jax.numpy as jnp # noqa: F811 _A = jnp.array _A = _is_jax else: _A = np.asarray _A = _is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: _A = [inputs] if not is_tensor(lowerCAmelCase_ ): _A = as_tensor(lowerCAmelCase_ ) except: # noqa E722 raise ValueError( """Unable to create tensor, you should probably activate truncation and/or padding """ """with 'padding=True' 'truncation=True' to have batched tensors with the same length.""" ) return inputs def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="" , lowerCAmelCase_="pt" ) -> BatchEncoding: _A = [0, 0, 0] _A = [artist] * len(self.version ) _A = [genres] * len(self.version ) _A , _A , _A = self.tokenize(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A , _A , _A = self._convert_token_to_id(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = [-INFINITY] * len(full_tokens[-1] ) _A = [ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]] , tensor_type=lowerCAmelCase_ ) for i in range(len(self.version ) ) ] return BatchEncoding({"""input_ids""": input_ids, """attention_masks""": attention_masks} ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""artists_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.artists_encoder , ensure_ascii=lowerCAmelCase_ ) ) _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""genres_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.genres_encoder , ensure_ascii=lowerCAmelCase_ ) ) _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""lyrics_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.lyrics_encoder , ensure_ascii=lowerCAmelCase_ ) ) return (artists_file, genres_file, lyrics_file) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[Any]: _A = self.artists_decoder.get(lowerCAmelCase_ ) _A = [self.genres_decoder.get(lowerCAmelCase_ ) for genre in genres_index] _A = [self.lyrics_decoder.get(lowerCAmelCase_ ) for character in lyric_index] return artist, genres, lyrics
83
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _SCREAMING_SNAKE_CASE = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def snake_case ( snake_case__ :Union[str, Any]) -> Dict: _A = torch.load(snake_case__ , map_location="""cpu""") return sd def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any] , snake_case__ :int=rename_keys_prefix) -> Optional[Any]: _A = OrderedDict() _A = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _A = key for name_pair in rename_keys_prefix: _A = new_key.replace(name_pair[0] , name_pair[1]) _A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _A = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple) -> int: assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _A = """pretraining""" if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "nlvr" in checkpoint_path: _A = {"""visual_embedding_dim""": 1_024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} _A = """multichoice""" elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} _A = """vqa_advanced""" elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048, """num_labels""": 3_129} _A = """vqa""" elif "nlvr" in checkpoint_path: _A = { """visual_embedding_dim""": 1_024, """num_labels""": 2, } _A = """nlvr""" _A = VisualBertConfig(**snake_case__) # Load State Dict _A = load_state_dict(snake_case__) _A = get_new_dict(snake_case__ , snake_case__) if model_type == "pretraining": _A = VisualBertForPreTraining(snake_case__) elif model_type == "vqa": _A = VisualBertForQuestionAnswering(snake_case__) elif model_type == "nlvr": _A = VisualBertForVisualReasoning(snake_case__) elif model_type == "multichoice": _A = VisualBertForMultipleChoice(snake_case__) model.load_state_dict(snake_case__) # Save Checkpoints Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
83
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''philschmid/bart-large-cnn-samsum''' lowerCamelCase :Tuple = ( '''This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ''' '''and returns a summary of the text.''' ) lowerCamelCase :List[Any] = '''summarizer''' lowerCamelCase :List[str] = AutoTokenizer lowerCamelCase :Dict = AutoModelForSeqaSeqLM lowerCamelCase :int = ['''text'''] lowerCamelCase :List[Any] = ['''text'''] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: return self.pre_processor(lowerCAmelCase_ , return_tensors="""pt""" , truncation=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: return self.model.generate(**lowerCAmelCase_ )[0] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: return self.pre_processor.decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ )
83
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def UpperCAmelCase ( self ) -> Optional[int]: _A = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCAmelCase_ ): self.assertDictEqual(lowerCAmelCase_ , example_records[i] ) def UpperCAmelCase ( self ) -> str: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) _A = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def UpperCAmelCase ( self ) -> Any: # checks what happens with missing columns _A = [{"""col_1""": 1}, {"""col_2""": """x"""}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def UpperCAmelCase ( self ) -> Tuple: # checks if the type can be inferred from the second record _A = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def UpperCAmelCase ( self ) -> Any: _A = Dataset.from_list([] ) self.assertEqual(len(lowerCAmelCase_ ) , 0 ) self.assertListEqual(dset.column_names , [] )
83
1
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger() def snake_case ( snake_case__ :int , snake_case__ :str , snake_case__ :LevitConfig , snake_case__ :Path , snake_case__ :bool = True) -> str: print(F'''Converting {name}...''') with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": _A = timm.create_model("""levit_128s""" , pretrained=snake_case__) else: _A = timm.create_model("""levit_128""" , pretrained=snake_case__) if hidden_sizes == 192: _A = timm.create_model("""levit_192""" , pretrained=snake_case__) if hidden_sizes == 256: _A = timm.create_model("""levit_256""" , pretrained=snake_case__) if hidden_sizes == 384: _A = timm.create_model("""levit_384""" , pretrained=snake_case__) from_model.eval() _A = LevitForImageClassificationWithTeacher(snake_case__).eval() _A = OrderedDict() _A = from_model.state_dict() _A = list(from_model.state_dict().keys()) _A = list(our_model.state_dict().keys()) print(len(snake_case__) , len(snake_case__)) for i in range(len(snake_case__)): _A = weights[og_keys[i]] our_model.load_state_dict(snake_case__) _A = torch.randn((2, 3, 224, 224)) _A = from_model(snake_case__) _A = our_model(snake_case__).logits assert torch.allclose(snake_case__ , snake_case__), "The model logits don't match the original one." _A = name print(snake_case__) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name) _A = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name) print(F'''Pushed {checkpoint_name}''') def snake_case ( snake_case__ :Path , snake_case__ :str = None , snake_case__ :bool = True) -> Optional[Any]: _A = """imagenet-1k-id2label.json""" _A = 1_000 _A = (1, num_labels) _A = """huggingface/label-files""" _A = num_labels _A = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""") , """r""")) _A = {int(snake_case__): v for k, v in idalabel.items()} _A = idalabel _A = {v: k for k, v in idalabel.items()} _A = partial(snake_case__ , num_labels=snake_case__ , idalabel=snake_case__ , labelaid=snake_case__) _A = { """levit-128S""": 128, """levit-128""": 128, """levit-192""": 192, """levit-256""": 256, """levit-384""": 384, } _A = { """levit-128S""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-128""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-192""": ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-256""": ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-384""": ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , snake_case__ , names_to_config[model_name] , snake_case__ , snake_case__) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , snake_case__ , snake_case__ , snake_case__ , snake_case__) return config, expected_shape if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help='The name of the model you wish to convert, it must be one of the supported Levit* architecture,', ) parser.add_argument( '--pytorch_dump_folder_path', default='levit-dump-folder/', type=Path, required=False, help='Path to the output PyTorch model directory.', ) parser.add_argument('--push_to_hub', action='store_true', help='Push model and image processor to the hub') parser.add_argument( '--no-push_to_hub', dest='push_to_hub', action='store_false', help='Do not push model and image processor to the hub', ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
83
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case__ , snake_case__))) _A = [float(snake_case__) for n in range(limit + 1)] for p in primes: for n in range(snake_case__ , limit + 1 , snake_case__): phi[n] *= 1 - 1 / p return int(sum(phi[2:])) if __name__ == "__main__": print(F'''{solution() = }''')
83
1
import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) _SCREAMING_SNAKE_CASE = 'bert-base-cased' _SCREAMING_SNAKE_CASE = 'fp16' _SCREAMING_SNAKE_CASE = 'bf16' _SCREAMING_SNAKE_CASE = [FPaa, BFaa] @require_fsdp @require_cuda class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: super().setUp() _A = dict( ACCELERATE_USE_FSDP="""true""" , MASTER_ADDR="""localhost""" , MASTER_PORT="""10999""" , RANK="""0""" , LOCAL_RANK="""0""" , WORLD_SIZE="""1""" , ) def UpperCAmelCase ( self ) -> int: from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(lowerCAmelCase_ ): _A = self.dist_env.copy() _A = F'''{i + 1}''' _A = strategy with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1 ) ) def UpperCAmelCase ( self ) -> Optional[Any]: from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(lowerCAmelCase_ ): _A = self.dist_env.copy() _A = prefetch_policy with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch ) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1 ) ) def UpperCAmelCase ( self ) -> Dict: from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(lowerCAmelCase_ ): _A = self.dist_env.copy() _A = state_dict_type with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1 ) ) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu ) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = AutoModel.from_pretrained(lowerCAmelCase_ ) for policy in FSDP_AUTO_WRAP_POLICY: _A = self.dist_env.copy() _A = policy if policy == "TRANSFORMER_BASED_WRAP": _A = """BertLayer""" elif policy == "SIZE_BASED_WRAP": _A = """2000""" with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(lowerCAmelCase_ ) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy ) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy ) _A = self.dist_env.copy() _A = """TRANSFORMER_BASED_WRAP""" _A = """T5Layer""" with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() with self.assertRaises(lowerCAmelCase_ ) as cm: fsdp_plugin.set_auto_wrap_policy(lowerCAmelCase_ ) self.assertTrue("""Could not find the transformer layer class to wrap in the model.""" in str(cm.exception ) ) _A = self.dist_env.copy() _A = """SIZE_BASED_WRAP""" _A = """0""" with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(lowerCAmelCase_ ) self.assertIsNone(fsdp_plugin.auto_wrap_policy ) def UpperCAmelCase ( self ) -> Optional[Any]: from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: _A = self.dist_env.copy() _A = mp_dtype with mockenv_context(**lowerCAmelCase_ ): _A = Accelerator() if mp_dtype == "fp16": _A = torch.floataa elif mp_dtype == "bf16": _A = torch.bfloataa _A = MixedPrecision(param_dtype=lowerCAmelCase_ , reduce_dtype=lowerCAmelCase_ , buffer_dtype=lowerCAmelCase_ ) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , lowerCAmelCase_ ) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , lowerCAmelCase_ ) ) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler ) AcceleratorState._reset_state(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: _A = self.dist_env.copy() _A = str(lowerCAmelCase_ ).lower() with mockenv_context(**lowerCAmelCase_ ): _A = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=lowerCAmelCase_ ) ) @require_fsdp @require_multi_gpu @slow class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: super().setUp() _A = 0.82 _A = [ """fsdp_shard_grad_op_transformer_based_wrap""", """fsdp_full_shard_transformer_based_wrap""", ] _A = { """multi_gpu_fp16""": 32_00, """fsdp_shard_grad_op_transformer_based_wrap_fp16""": 20_00, """fsdp_full_shard_transformer_based_wrap_fp16""": 19_00, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } _A = 1_60 _A = 1_60 _A = inspect.getfile(accelerate.test_utils ) _A = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """external_deps"""] ) def UpperCAmelCase ( self ) -> List[str]: _A = os.path.join(self.test_scripts_folder , """test_performance.py""" ) _A = ["""accelerate""", """launch""", """--num_processes=2""", """--num_machines=1""", """--machine_rank=0""", """--use_fsdp"""] for config in self.performance_configs: _A = cmd.copy() for i, strategy in enumerate(lowerCAmelCase_ ): if strategy.lower() in config: cmd_config.append(F'''--fsdp_sharding_strategy={i+1}''' ) break if "fp32" in config: cmd_config.append("""--mixed_precision=no""" ) else: cmd_config.append("""--mixed_precision=fp16""" ) if "cpu_offload" in config: cmd_config.append("""--fsdp_offload_params=True""" ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(F'''--fsdp_auto_wrap_policy={policy}''' ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("""--fsdp_transformer_layer_cls_to_wrap=BertLayer""" ) elif policy == "SIZE_BASED_WRAP": cmd_config.append("""--fsdp_min_num_params=2000""" ) cmd_config.extend( [ self.test_file_path, F'''--output_dir={self.tmpdir}''', F'''--performance_lower_bound={self.performance_lower_bound}''', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() ) def UpperCAmelCase ( self ) -> List[str]: _A = os.path.join(self.test_scripts_folder , """test_checkpointing.py""" ) _A = [ """accelerate""", """launch""", """--num_processes=2""", """--num_machines=1""", """--machine_rank=0""", """--use_fsdp""", """--mixed_precision=fp16""", """--fsdp_transformer_layer_cls_to_wrap=BertLayer""", ] for i, strategy in enumerate(lowerCAmelCase_ ): _A = cmd.copy() cmd_config.append(F'''--fsdp_sharding_strategy={i+1}''' ) if strategy != "FULL_SHARD": continue _A = len(lowerCAmelCase_ ) for state_dict_type in FSDP_STATE_DICT_TYPE: _A = cmd_config[:state_dict_config_index] cmd_config.append(F'''--fsdp_state_dict_type={state_dict_type}''' ) cmd_config.extend( [ self.test_file_path, F'''--output_dir={self.tmpdir}''', """--partial_train_epoch=1""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() ) _A = cmd_config[:-1] _A = os.path.join(self.tmpdir , """epoch_0""" ) cmd_config.extend( [ F'''--resume_from_checkpoint={resume_from_checkpoint}''', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() ) def UpperCAmelCase ( self ) -> Any: _A = os.path.join(self.test_scripts_folder , """test_peak_memory_usage.py""" ) _A = [ """accelerate""", """launch""", """--num_processes=2""", """--num_machines=1""", """--machine_rank=0""", ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): _A = cmd.copy() if "fp16" in spec: cmd_config.extend(["""--mixed_precision=fp16"""] ) else: cmd_config.extend(["""--mixed_precision=no"""] ) if "multi_gpu" in spec: continue else: cmd_config.extend(["""--use_fsdp"""] ) for i, strategy in enumerate(lowerCAmelCase_ ): if strategy.lower() in spec: cmd_config.append(F'''--fsdp_sharding_strategy={i+1}''' ) break if "cpu_offload" in spec: cmd_config.append("""--fsdp_offload_params=True""" ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(F'''--fsdp_auto_wrap_policy={policy}''' ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("""--fsdp_transformer_layer_cls_to_wrap=BertLayer""" ) elif policy == "SIZE_BASED_WRAP": cmd_config.append("""--fsdp_min_num_params=2000""" ) cmd_config.extend( [ self.test_file_path, F'''--output_dir={self.tmpdir}''', F'''--peak_memory_upper_bound={peak_mem_upper_bound}''', F'''--n_train={self.n_train}''', F'''--n_val={self.n_val}''', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() )
83
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=32 , lowerCAmelCase_=5 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_="None" , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=None , ) -> Union[str, Any]: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = num_choices _A = relative_attention _A = position_biased_input _A = pos_att_type _A = scope def UpperCAmelCase ( self ) -> Dict: _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self ) -> Optional[int]: return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Any: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: _A = DebertaVaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = DebertaVaForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: _A = self.num_labels _A = DebertaVaForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = self.num_labels _A = DebertaVaForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _A = DebertaVaForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = DebertaVaForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) lowerCamelCase :str = ( { '''feature-extraction''': DebertaVaModel, '''fill-mask''': DebertaVaForMaskedLM, '''question-answering''': DebertaVaForQuestionAnswering, '''text-classification''': DebertaVaForSequenceClassification, '''token-classification''': DebertaVaForTokenClassification, '''zero-shot''': DebertaVaForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase :str = True lowerCamelCase :Union[str, Any] = False lowerCamelCase :Optional[int] = False lowerCamelCase :List[str] = False lowerCamelCase :str = False def UpperCAmelCase ( self ) -> Optional[int]: _A = DebertaVaModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> List[str]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> List[str]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> Any: for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = DebertaVaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason="""Model not available yet""" ) def UpperCAmelCase ( self ) -> int: pass @slow def UpperCAmelCase ( self ) -> Optional[Any]: _A = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _A = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) _A = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] # compare the actual values for a slice. _A = torch.tensor( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
83
1
import numpy as np import qiskit def snake_case ( snake_case__ :int = 8 , snake_case__ :int | None = None) -> str: _A = np.random.default_rng(seed=snake_case__) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _A = 6 * key_len # Measurement basis for Alice's qubits. _A = rng.integers(2 , size=snake_case__) # The set of states Alice will prepare. _A = rng.integers(2 , size=snake_case__) # Measurement basis for Bob's qubits. _A = rng.integers(2 , size=snake_case__) # Quantum Circuit to simulate BB84 _A = qiskit.QuantumCircuit(snake_case__ , name="""BB84""") # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case__): if alice_state[index] == 1: bbaa_circ.x(snake_case__) if alice_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case__): if bob_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _A = qiskit.Aer.get_backend("""aer_simulator""") # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _A = qiskit.execute(snake_case__ , snake_case__ , shots=1 , seed_simulator=snake_case__) # Returns the result of measurement. _A = job.result().get_counts(snake_case__).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _A = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case__ , snake_case__ , snake_case__) if alice_basis_bit == bob_basis_bit ]) # Get final key. Pad with 0 if too short, otherwise truncate. _A = gen_key[:key_len] if len(snake_case__) >= key_len else gen_key.ljust(snake_case__ , """0""") return key if __name__ == "__main__": print(F'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
83
def snake_case ( snake_case__ :int , snake_case__ :int) -> int: return int(input_a == input_a == 0) def snake_case ( ) -> None: print("""Truth Table of NOR Gate:""") print("""| Input 1 | Input 2 | Output |""") print(F'''| 0 | 0 | {nor_gate(0 , 0)} |''') print(F'''| 0 | 1 | {nor_gate(0 , 1)} |''') print(F'''| 1 | 0 | {nor_gate(1 , 0)} |''') print(F'''| 1 | 1 | {nor_gate(1 , 1)} |''') if __name__ == "__main__": import doctest doctest.testmod() main()
83
1
import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _SCREAMING_SNAKE_CASE = [ # (stable-diffusion, HF Diffusers) ('time_embed.0.weight', 'time_embedding.linear_1.weight'), ('time_embed.0.bias', 'time_embedding.linear_1.bias'), ('time_embed.2.weight', 'time_embedding.linear_2.weight'), ('time_embed.2.bias', 'time_embedding.linear_2.bias'), ('input_blocks.0.0.weight', 'conv_in.weight'), ('input_blocks.0.0.bias', 'conv_in.bias'), ('out.0.weight', 'conv_norm_out.weight'), ('out.0.bias', 'conv_norm_out.bias'), ('out.2.weight', 'conv_out.weight'), ('out.2.bias', 'conv_out.bias'), ] _SCREAMING_SNAKE_CASE = [ # (stable-diffusion, HF Diffusers) ('in_layers.0', 'norm1'), ('in_layers.2', 'conv1'), ('out_layers.0', 'norm2'), ('out_layers.3', 'conv2'), ('emb_layers.1', 'time_emb_proj'), ('skip_connection', 'conv_shortcut'), ] _SCREAMING_SNAKE_CASE = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _SCREAMING_SNAKE_CASE = F'''down_blocks.{i}.resnets.{j}.''' _SCREAMING_SNAKE_CASE = F'''input_blocks.{3*i + j + 1}.0.''' unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _SCREAMING_SNAKE_CASE = F'''down_blocks.{i}.attentions.{j}.''' _SCREAMING_SNAKE_CASE = F'''input_blocks.{3*i + j + 1}.1.''' unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _SCREAMING_SNAKE_CASE = F'''up_blocks.{i}.resnets.{j}.''' _SCREAMING_SNAKE_CASE = F'''output_blocks.{3*i + j}.0.''' unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _SCREAMING_SNAKE_CASE = F'''up_blocks.{i}.attentions.{j}.''' _SCREAMING_SNAKE_CASE = F'''output_blocks.{3*i + j}.1.''' unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _SCREAMING_SNAKE_CASE = F'''down_blocks.{i}.downsamplers.0.conv.''' _SCREAMING_SNAKE_CASE = F'''input_blocks.{3*(i+1)}.0.op.''' unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _SCREAMING_SNAKE_CASE = F'''up_blocks.{i}.upsamplers.0.''' _SCREAMING_SNAKE_CASE = F'''output_blocks.{3*i + 2}.{1 if i == 0 else 2}.''' unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _SCREAMING_SNAKE_CASE = 'mid_block.attentions.0.' _SCREAMING_SNAKE_CASE = 'middle_block.1.' unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _SCREAMING_SNAKE_CASE = F'''mid_block.resnets.{j}.''' _SCREAMING_SNAKE_CASE = F'''middle_block.{2*j}.''' unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def snake_case ( snake_case__ :List[str]) -> int: # buyer beware: this is a *brittle* function, # and correct output requires that all of these pieces interact in # the exact order in which I have arranged them. _A = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _A = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _A = v.replace(snake_case__ , snake_case__) _A = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _A = v.replace(snake_case__ , snake_case__) _A = v _A = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _SCREAMING_SNAKE_CASE = [ # (stable-diffusion, HF Diffusers) ('nin_shortcut', 'conv_shortcut'), ('norm_out', 'conv_norm_out'), ('mid.attn_1.', 'mid_block.attentions.0.'), ] for i in range(4): # down_blocks have two resnets for j in range(2): _SCREAMING_SNAKE_CASE = F'''encoder.down_blocks.{i}.resnets.{j}.''' _SCREAMING_SNAKE_CASE = F'''encoder.down.{i}.block.{j}.''' vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _SCREAMING_SNAKE_CASE = F'''down_blocks.{i}.downsamplers.0.''' _SCREAMING_SNAKE_CASE = F'''down.{i}.downsample.''' vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _SCREAMING_SNAKE_CASE = F'''up_blocks.{i}.upsamplers.0.''' _SCREAMING_SNAKE_CASE = F'''up.{3-i}.upsample.''' vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _SCREAMING_SNAKE_CASE = F'''decoder.up_blocks.{i}.resnets.{j}.''' _SCREAMING_SNAKE_CASE = F'''decoder.up.{3-i}.block.{j}.''' vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _SCREAMING_SNAKE_CASE = F'''mid_block.resnets.{i}.''' _SCREAMING_SNAKE_CASE = F'''mid.block_{i+1}.''' vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _SCREAMING_SNAKE_CASE = [ # (stable-diffusion, HF Diffusers) ('norm.', 'group_norm.'), ('q.', 'query.'), ('k.', 'key.'), ('v.', 'value.'), ('proj_out.', 'proj_attn.'), ] def snake_case ( snake_case__ :List[str]) -> List[Any]: # convert HF linear weights to SD conv2d weights return w.reshape(*w.shape , 1 , 1) def snake_case ( snake_case__ :List[str]) -> Tuple: _A = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _A = v.replace(snake_case__ , snake_case__) _A = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _A = v.replace(snake_case__ , snake_case__) _A = v _A = {v: vae_state_dict[k] for k, v in mapping.items()} _A = ["""q""", """k""", """v""", """proj_out"""] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''') _A = reshape_weight_for_sd(snake_case__) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _SCREAMING_SNAKE_CASE = [ # (stable-diffusion, HF Diffusers) ('resblocks.', 'text_model.encoder.layers.'), ('ln_1', 'layer_norm1'), ('ln_2', 'layer_norm2'), ('.c_fc.', '.fc1.'), ('.c_proj.', '.fc2.'), ('.attn', '.self_attn'), ('ln_final.', 'transformer.text_model.final_layer_norm.'), ('token_embedding.weight', 'transformer.text_model.embeddings.token_embedding.weight'), ('positional_embedding', 'transformer.text_model.embeddings.position_embedding.weight'), ] _SCREAMING_SNAKE_CASE = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _SCREAMING_SNAKE_CASE = re.compile('|'.join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _SCREAMING_SNAKE_CASE = {'q': 0, 'k': 1, 'v': 2} def snake_case ( snake_case__ :Optional[int]) -> Tuple: _A = {} _A = {} _A = {} for k, v in text_enc_dict.items(): if ( k.endswith(""".self_attn.q_proj.weight""") or k.endswith(""".self_attn.k_proj.weight""") or k.endswith(""".self_attn.v_proj.weight""") ): _A = k[: -len(""".q_proj.weight""")] _A = k[-len("""q_proj.weight""")] if k_pre not in capture_qkv_weight: _A = [None, None, None] _A = v continue if ( k.endswith(""".self_attn.q_proj.bias""") or k.endswith(""".self_attn.k_proj.bias""") or k.endswith(""".self_attn.v_proj.bias""") ): _A = k[: -len(""".q_proj.bias""")] _A = k[-len("""q_proj.bias""")] if k_pre not in capture_qkv_bias: _A = [None, None, None] _A = v continue _A = textenc_pattern.sub(lambda snake_case__: protected[re.escape(m.group(0))] , snake_case__) _A = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception("""CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing""") _A = textenc_pattern.sub(lambda snake_case__: protected[re.escape(m.group(0))] , snake_case__) _A = torch.cat(snake_case__) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception("""CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing""") _A = textenc_pattern.sub(lambda snake_case__: protected[re.escape(m.group(0))] , snake_case__) _A = torch.cat(snake_case__) return new_state_dict def snake_case ( snake_case__ :Optional[Any]) -> List[str]: return text_enc_dict if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--half', action='store_true', help='Save weights in half precision.') parser.add_argument( '--use_safetensors', action='store_true', help='Save weights use safetensors, default is ckpt.' ) _SCREAMING_SNAKE_CASE = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.safetensors') _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.safetensors') _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'text_encoder', 'model.safetensors') # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _SCREAMING_SNAKE_CASE = load_file(unet_path, device='cpu') else: _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.bin') _SCREAMING_SNAKE_CASE = torch.load(unet_path, map_location='cpu') if osp.exists(vae_path): _SCREAMING_SNAKE_CASE = load_file(vae_path, device='cpu') else: _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.bin') _SCREAMING_SNAKE_CASE = torch.load(vae_path, map_location='cpu') if osp.exists(text_enc_path): _SCREAMING_SNAKE_CASE = load_file(text_enc_path, device='cpu') else: _SCREAMING_SNAKE_CASE = osp.join(args.model_path, 'text_encoder', 'pytorch_model.bin') _SCREAMING_SNAKE_CASE = torch.load(text_enc_path, map_location='cpu') # Convert the UNet model _SCREAMING_SNAKE_CASE = convert_unet_state_dict(unet_state_dict) _SCREAMING_SNAKE_CASE = {'model.diffusion_model.' + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _SCREAMING_SNAKE_CASE = convert_vae_state_dict(vae_state_dict) _SCREAMING_SNAKE_CASE = {'first_stage_model.' + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _SCREAMING_SNAKE_CASE = 'text_model.encoder.layers.22.layer_norm2.bias' in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _SCREAMING_SNAKE_CASE = {'transformer.' + k: v for k, v in text_enc_dict.items()} _SCREAMING_SNAKE_CASE = convert_text_enc_state_dict_vaa(text_enc_dict) _SCREAMING_SNAKE_CASE = {'cond_stage_model.model.' + k: v for k, v in text_enc_dict.items()} else: _SCREAMING_SNAKE_CASE = convert_text_enc_state_dict(text_enc_dict) _SCREAMING_SNAKE_CASE = {'cond_stage_model.transformer.' + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _SCREAMING_SNAKE_CASE = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _SCREAMING_SNAKE_CASE = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _SCREAMING_SNAKE_CASE = {'state_dict': state_dict} torch.save(state_dict, args.checkpoint_path)
83
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=sys.maxsize ) -> str: _A = """bilinear""" _A = max_size _A = short_edge_length def __call__( self , lowerCAmelCase_ ) -> Optional[Any]: _A = [] for img in imgs: _A , _A = img.shape[:2] # later: provide list and randomly choose index for resize _A = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img _A = size * 1.0 / min(lowerCAmelCase_ , lowerCAmelCase_ ) if h < w: _A , _A = size, scale * w else: _A , _A = scale * h, size if max(lowerCAmelCase_ , lowerCAmelCase_ ) > self.max_size: _A = self.max_size * 1.0 / max(lowerCAmelCase_ , lowerCAmelCase_ ) _A = newh * scale _A = neww * scale _A = int(neww + 0.5 ) _A = int(newh + 0.5 ) if img.dtype == np.uinta: _A = Image.fromarray(lowerCAmelCase_ ) _A = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) _A = np.asarray(lowerCAmelCase_ ) else: _A = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw _A = nn.functional.interpolate( lowerCAmelCase_ , (newh, neww) , mode=self.interp_method , align_corners=lowerCAmelCase_ ).squeeze(0 ) img_augs.append(lowerCAmelCase_ ) return img_augs class a : """simple docstring""" def __init__( self , lowerCAmelCase_ ) -> List[Any]: _A = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) _A = cfg.INPUT.FORMAT _A = cfg.SIZE_DIVISIBILITY _A = cfg.PAD_VALUE _A = cfg.INPUT.MAX_SIZE_TEST _A = cfg.MODEL.DEVICE _A = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = lambda lowerCAmelCase_ : (x - self.pixel_mean) / self.pixel_std def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = tuple(max(lowerCAmelCase_ ) for s in zip(*[img.shape for img in images] ) ) _A = [im.shape[-2:] for im in images] _A = [ nn.functional.pad( lowerCAmelCase_ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] return torch.stack(lowerCAmelCase_ ), torch.tensor(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_=False ) -> int: with torch.no_grad(): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = [images] if single_image: assert len(lowerCAmelCase_ ) == 1 for i in range(len(lowerCAmelCase_ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(lowerCAmelCase_ , images.pop(lowerCAmelCase_ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( lowerCAmelCase_ , torch.as_tensor(img_tensorize(images.pop(lowerCAmelCase_ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge _A = torch.tensor([im.shape[:2] for im in images] ) _A = self.aug(lowerCAmelCase_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic _A = [self.normalizer(lowerCAmelCase_ ) for x in images] # now pad them to do the following operations _A , _A = self.pad(lowerCAmelCase_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad _A = torch.true_divide(lowerCAmelCase_ , lowerCAmelCase_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[Any]) -> Tuple: boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def snake_case ( snake_case__ :Optional[int] , snake_case__ :Tuple[int, int]) -> Optional[Any]: assert torch.isfinite(snake_case__).all(), "Box tensor contains infinite or NaN!" _A , _A = box_size tensor[:, 0].clamp_(min=0 , max=snake_case__) tensor[:, 1].clamp_(min=0 , max=snake_case__) tensor[:, 2].clamp_(min=0 , max=snake_case__) tensor[:, 3].clamp_(min=0 , max=snake_case__)
83
1
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=3 , lowerCAmelCase_=32 , lowerCAmelCase_=3 , lowerCAmelCase_=10 , lowerCAmelCase_=[10, 20, 30, 40] , lowerCAmelCase_=[1, 1, 2, 1] , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=3 , lowerCAmelCase_=None , ) -> Optional[Any]: _A = parent _A = batch_size _A = image_size _A = num_channels _A = embeddings_size _A = hidden_sizes _A = depths _A = is_training _A = use_labels _A = hidden_act _A = num_labels _A = scope _A = len(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.num_labels ) _A = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self ) -> Union[str, Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[Any]: _A = RegNetModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _A = self.num_labels _A = RegNetForImageClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self ) -> Any: _A = self.prepare_config_and_inputs() _A , _A , _A = config_and_inputs _A = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :str = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () lowerCamelCase :Dict = ( {'''feature-extraction''': RegNetModel, '''image-classification''': RegNetForImageClassification} if is_torch_available() else {} ) lowerCamelCase :str = False lowerCamelCase :Optional[Any] = False lowerCamelCase :Any = False lowerCamelCase :Tuple = False def UpperCAmelCase ( self ) -> Any: _A = RegNetModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self ) -> Dict: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase ( self ) -> Optional[Any]: pass @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase ( self ) -> Optional[int]: pass def UpperCAmelCase ( self ) -> Union[str, Any]: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(lowerCAmelCase_ ) _A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A = [*signature.parameters.keys()] _A = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(config=lowerCAmelCase_ ) for name, module in model.named_modules(): if isinstance(lowerCAmelCase_ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) def UpperCAmelCase ( self ) -> List[Any]: def check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): _A = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) ) _A = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _A = self.model_tester.num_stages self.assertEqual(len(lowerCAmelCase_ ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) _A , _A = self.model_tester.prepare_config_and_inputs_for_common() _A = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: _A = layer_type _A = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _A = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> Any: for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = RegNetModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def snake_case ( ) -> Optional[int]: _A = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""") return image @require_torch @require_vision class a ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase ( self ) -> Dict: return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase ( self ) -> Union[str, Any]: _A = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCAmelCase_ ) _A = self.default_image_processor _A = prepare_img() _A = image_processor(images=lowerCAmelCase_ , return_tensors="""pt""" ).to(lowerCAmelCase_ ) # forward pass with torch.no_grad(): _A = model(**lowerCAmelCase_ ) # verify the logits _A = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) _A = torch.tensor([-0.4180, -1.5051, -3.4836] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase_ , atol=1E-4 ) )
83
from collections import defaultdict def snake_case ( snake_case__ :int) -> int: _A = 1 _A = True for v in tree[start]: if v not in visited: ret += dfs(snake_case__) if ret % 2 == 0: cuts.append(snake_case__) return ret def snake_case ( ) -> Any: dfs(1) if __name__ == "__main__": _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10, 9 _SCREAMING_SNAKE_CASE = defaultdict(list) _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
83
1
import colorsys from PIL import Image # type: ignore def snake_case ( snake_case__ :float , snake_case__ :float , snake_case__ :int) -> float: _A = x _A = y for step in range(snake_case__): # noqa: B007 _A = a * a - b * b + x _A = 2 * a * b + y _A = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255) for i in colorsys.hsv_to_rgb(snake_case__ , 1 , 1)) def snake_case ( snake_case__ :int = 800 , snake_case__ :int = 600 , snake_case__ :float = -0.6 , snake_case__ :float = 0 , snake_case__ :float = 3.2 , snake_case__ :int = 50 , snake_case__ :bool = True , ) -> Image.Image: _A = Image.new("""RGB""" , (image_width, image_height)) _A = img.load() # loop through the image-coordinates for image_x in range(snake_case__): for image_y in range(snake_case__): # determine the figure-coordinates based on the image-coordinates _A = figure_width / image_width * image_height _A = figure_center_x + (image_x / image_width - 0.5) * figure_width _A = figure_center_y + (image_y / image_height - 0.5) * figure_height _A = get_distance(snake_case__ , snake_case__ , snake_case__) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _A = get_color_coded_rgb(snake_case__) else: _A = get_black_and_white_rgb(snake_case__) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _SCREAMING_SNAKE_CASE = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
import heapq def snake_case ( snake_case__ :dict) -> set[int]: _A = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(snake_case__ , [-1 * len(snake_case__), (key, value)]) # chosen_vertices = set of chosen vertices _A = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _A = heapq.heappop(snake_case__)[1][0] chosen_vertices.add(snake_case__) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _A = elem[1][1].index(snake_case__) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(snake_case__) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(F'''Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}''')
83
1
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def snake_case ( snake_case__ :Union[str, Any]) -> str: return {key.lstrip("""-"""): value for key, value in zip(unknown_args[::2] , unknown_args[1::2])} def snake_case ( ) -> str: _A = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=snake_case__) _A = parser.add_subparsers(help="""datasets-cli command helpers""") set_verbosity_info() # Register commands ConvertCommand.register_subcommand(snake_case__) EnvironmentCommand.register_subcommand(snake_case__) TestCommand.register_subcommand(snake_case__) RunBeamCommand.register_subcommand(snake_case__) DummyDataCommand.register_subcommand(snake_case__) # Parse args _A , _A = parser.parse_known_args() if not hasattr(snake_case__ , """func"""): parser.print_help() exit(1) _A = parse_unknown_args(snake_case__) # Run _A = args.func(snake_case__ , **snake_case__) service.run() if __name__ == "__main__": main()
83
import math import unittest def snake_case ( snake_case__ :int) -> bool: assert isinstance(snake_case__ , snake_case__) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def UpperCAmelCase ( self ) -> Dict: with self.assertRaises(lowerCAmelCase_ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , """Zero doesn't have any positive factors, primes must have exactly two.""" , ) self.assertFalse( is_prime(1 ) , """One only has 1 positive factor, primes must have exactly two.""" , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
83
1
# HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers _SCREAMING_SNAKE_CASE = float('nan') class a : """simple docstring""" def __init__( self , lowerCAmelCase_ ) -> Any: _A = sys.stdout _A = open(lowerCAmelCase_ , """a""" ) def __getattr__( self , lowerCAmelCase_ ) -> str: return getattr(self.stdout , lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Optional[Any]: self.stdout.write(lowerCAmelCase_ ) # strip tqdm codes self.file.write(re.sub(r"""^.*\r""" , """""" , lowerCAmelCase_ , 0 , re.M ) ) def snake_case ( snake_case__ :Optional[Any]=80 , snake_case__ :Optional[Any]=False) -> Dict: _A = [] # deal with critical env vars _A = ["""CUDA_VISIBLE_DEVICES"""] for key in env_keys: _A = os.environ.get(snake_case__ , snake_case__) if val is not None: cmd.append(F'''{key}={val}''') # python executable (not always needed if the script is executable) _A = sys.executable if full_python_path else sys.executable.split("""/""")[-1] cmd.append(snake_case__) # now the normal args cmd += list(map(shlex.quote , sys.argv)) # split up into up to MAX_WIDTH lines with shell multi-line escapes _A = [] _A = """""" while len(snake_case__) > 0: current_line += F'''{cmd.pop(0)} ''' if len(snake_case__) == 0 or len(snake_case__) + len(cmd[0]) + 1 > max_width - 1: lines.append(snake_case__) _A = """""" return "\\\n".join(snake_case__) def snake_case ( snake_case__ :int , snake_case__ :str) -> Any: # unwrap multi-line input _A = re.sub(R"""[\\\n]+""" , """ """ , args.base_cmd) # remove --output_dir if any and set our own _A = re.sub("""--output_dir\s+[^\s]+""" , """""" , args.base_cmd) args.base_cmd += F''' --output_dir {output_dir}''' # ensure we have --overwrite_output_dir _A = re.sub("""--overwrite_output_dir\s+""" , """""" , args.base_cmd) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd) def snake_case ( snake_case__ :int , snake_case__ :Optional[Any] , snake_case__ :str , snake_case__ :List[Any] , snake_case__ :Optional[int] , snake_case__ :Any , snake_case__ :Optional[int]) -> List[Any]: # Enable to debug everything but the run itself, to do it fast and see the progress. # This is useful for debugging the output formatting quickly - we can remove it later once # everybody is happy with the output if 0: import random from time import sleep sleep(0) return dict( {k: random.uniform(0 , 100) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6666, 222.2222_2222])} , ) _A = subprocess.run(snake_case__ , capture_output=snake_case__ , text=snake_case__) if verbose: print("""STDOUT""" , result.stdout) print("""STDERR""" , result.stderr) # save the streams _A = variation.replace(""" """ , """-""") with open(Path(snake_case__) / F'''log.{prefix}.stdout.txt''' , """w""") as f: f.write(result.stdout) with open(Path(snake_case__) / F'''log.{prefix}.stderr.txt''' , """w""") as f: f.write(result.stderr) if result.returncode != 0: if verbose: print("""failed""") return {target_metric_key: nan} with io.open(F'''{output_dir}/all_results.json''' , """r""" , encoding="""utf-8""") as f: _A = json.load(snake_case__) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def snake_case ( snake_case__ :str , snake_case__ :Tuple , snake_case__ :List[Any] , snake_case__ :List[str] , snake_case__ :Any , snake_case__ :List[str] , snake_case__ :Any , snake_case__ :Optional[Any] , snake_case__ :Union[str, Any] , snake_case__ :Tuple , ) -> Tuple: _A = [] _A = [] _A = F'''{id}: {variation:<{longest_variation_len}}''' _A = F'''{preamble}: ''' _A = set(report_metric_keys + [target_metric_key]) for i in tqdm(range(snake_case__) , desc=snake_case__ , leave=snake_case__): _A = process_run_single( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) _A = single_run_metrics[target_metric_key] if not math.isnan(snake_case__): metrics.append(snake_case__) results.append(snake_case__) outcome += "✓" else: outcome += "✘" _A = F'''\33[2K\r{outcome}''' if len(snake_case__) > 0: _A = {k: fmean([x[k] for x in metrics]) for k in metrics[0].keys()} _A = round(mean_metrics[target_metric_key] , 2) _A = F'''{outcome} {mean_target}''' if len(snake_case__) > 1: results_str += F''' {tuple(round(snake_case__ , 2) for x in results)}''' print(snake_case__) _A = variation return mean_metrics else: print(snake_case__) return {variation_key: variation, target_metric_key: nan} def snake_case ( ) -> Any: _A = torch.cuda.get_device_properties(torch.device("""cuda""")) return F''' Datetime : {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')} Software: transformers: {transformers.__version__} torch : {torch.__version__} cuda : {torch.version.cuda} python : {platform.python_version()} Hardware: {torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB ''' def snake_case ( snake_case__ :List[Any] , snake_case__ :int , snake_case__ :Optional[int] , snake_case__ :Tuple , snake_case__ :List[str]) -> Any: _A = pd.DataFrame(snake_case__) _A = """variation""" _A = """diff_%""" _A = nan if base_variation is not None and len(df[df[variation_key] == base_variation]): # this may still return nan _A = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(snake_case__): # as a fallback, use the minimal value as the sentinel _A = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(snake_case__): _A = df.apply( lambda snake_case__: round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value) if not math.isnan(r[target_metric_key]) else 0 , axis="""columns""" , ) # re-order columns _A = [variation_key, target_metric_key, diff_key, *report_metric_keys] _A = df.reindex(snake_case__ , axis="""columns""") # reorder cols # capitalize _A = df.rename(str.capitalize , axis="""columns""") # make the cols as narrow as possible _A = df.rename(lambda snake_case__: c.replace("""_""" , """<br>""") , axis="""columns""") _A = df.rename(lambda snake_case__: c.replace("""_""" , """\n""") , axis="""columns""") _A = ["""""", """Copy between the cut-here-lines and paste as is to github or a forum"""] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=snake_case__ , floatfmt=""".2f""")] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=snake_case__ , floatfmt=""".2f""")] print("""\n\n""".join(snake_case__)) def snake_case ( ) -> Any: _A = argparse.ArgumentParser() parser.add_argument( """--base-cmd""" , default=snake_case__ , type=snake_case__ , required=snake_case__ , help="""Base cmd""" , ) parser.add_argument( """--variations""" , default=snake_case__ , type=snake_case__ , nargs="""+""" , required=snake_case__ , help="""Multi-dimensional variations, example: '|--fp16|--bf16' '|--tf32'""" , ) parser.add_argument( """--base-variation""" , default=snake_case__ , type=snake_case__ , help="""Baseline variation to compare to. if None the minimal target value will be used to compare against""" , ) parser.add_argument( """--target-metric-key""" , default=snake_case__ , type=snake_case__ , required=snake_case__ , help="""Target metric key in output_dir/all_results.json, e.g., train_samples_per_second""" , ) parser.add_argument( """--report-metric-keys""" , default="""""" , type=snake_case__ , help="""Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., 'train_loss train_samples""" , ) parser.add_argument( """--repeat-times""" , default=1 , type=snake_case__ , help="""How many times to re-run each variation - an average will be reported""" , ) parser.add_argument( """--output_dir""" , default="""output_benchmark""" , type=snake_case__ , help="""The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked""" , ) parser.add_argument( """--verbose""" , default=snake_case__ , action="""store_true""" , help="""Whether to show the outputs of each run or just the benchmark progress""" , ) _A = parser.parse_args() _A = args.output_dir Path(snake_case__).mkdir(exist_ok=snake_case__) _A = get_base_command(snake_case__ , snake_case__) # split each dimension into its --foo variations _A = [list(map(str.strip , re.split(R"""\|""" , snake_case__))) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty _A = list(map(str.strip , map(""" """.join , itertools.product(*snake_case__)))) _A = max(len(snake_case__) for x in variations) # split wanted keys _A = args.report_metric_keys.split() # capture prints into a log file for convenience _A = F'''benchmark-report-{datetime.datetime.now().strftime('%Y-%m-%d-%H-%M-%S')}.txt''' print(F'''\nNote: each run\'s output is also logged under {output_dir}/log.*.std*.txt''') print(F'''and this script\'s output is also piped into {report_fn}''') _A = Tee(snake_case__) print(F'''\n*** Running {len(snake_case__)} benchmarks:''') print(F'''Base command: {' '.join(snake_case__)}''') _A = """variation""" _A = [] for id, variation in enumerate(tqdm(snake_case__ , desc="""Total completion: """ , leave=snake_case__)): _A = base_cmd + variation.split() results.append( process_run( id + 1 , snake_case__ , snake_case__ , snake_case__ , snake_case__ , args.target_metric_key , snake_case__ , args.repeat_times , snake_case__ , args.verbose , )) process_results(snake_case__ , args.target_metric_key , snake_case__ , args.base_variation , snake_case__) if __name__ == "__main__": main()
83
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers _SCREAMING_SNAKE_CASE = 'python tqdm regex requests packaging filelock numpy tokenizers'.split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append('dataclasses') if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append('importlib_metadata') for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def snake_case ( snake_case__ :int , snake_case__ :str=None) -> Any: require_version(deps[pkg] , snake_case__)
83
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(__lowerCAmelCase ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Optional[Any]: super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) self.check_model_type(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , **lowerCAmelCase_ ) -> Tuple: _A , _A = {}, {} if padding is not None: _A = padding if truncation is not None: _A = truncation if top_k is not None: _A = top_k return preprocess_params, {}, postprocess_params def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , **lowerCAmelCase_ ) -> Union[str, Any]: if isinstance(lowerCAmelCase_ , (Image.Image, str) ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = {"""image""": image, """question""": question} else: _A = image _A = super().__call__(lowerCAmelCase_ , **lowerCAmelCase_ ) return results def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Any: _A = load_image(inputs["""image"""] ) _A = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ ) _A = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase_ ) return model_inputs def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = self.model(**lowerCAmelCase_ ) return model_outputs def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=5 ) -> Union[str, Any]: if top_k > self.model.config.num_labels: _A = self.model.config.num_labels if self.framework == "pt": _A = model_outputs.logits.sigmoid()[0] _A , _A = probs.topk(lowerCAmelCase_ ) else: raise ValueError(F'''Unsupported framework: {self.framework}''' ) _A = scores.tolist() _A = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowerCAmelCase_ , lowerCAmelCase_ )]
83
1
import argparse from collections import defaultdict import yaml _SCREAMING_SNAKE_CASE = 'docs/source/en/_toctree.yml' def snake_case ( snake_case__ :int) -> Union[str, Any]: _A = defaultdict(snake_case__) for doc in model_doc: counts[doc["local"]] += 1 _A = [key for key, value in counts.items() if value > 1] _A = [] for duplicate_key in duplicates: _A = list({doc["""title"""] for doc in model_doc if doc["""local"""] == duplicate_key}) if len(snake_case__) > 1: raise ValueError( F'''{duplicate_key} is present several times in the documentation table of content at ''' """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""") # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]}) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc["""local"""]] == 1]) # Sort return sorted(snake_case__ , key=lambda snake_case__: s["title"].lower()) def snake_case ( snake_case__ :Tuple=False) -> Optional[Any]: with open(snake_case__ , encoding="""utf-8""") as f: _A = yaml.safe_load(f.read()) # Get to the API doc _A = 0 while content[api_idx]["title"] != "API": api_idx += 1 _A = content[api_idx]["""sections"""] # Then to the model doc _A = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 _A = api_doc[model_idx]["""sections"""] _A = [(idx, section) for idx, section in enumerate(snake_case__) if """sections""" in section] _A = False for idx, modality_doc in modalities_docs: _A = modality_doc["""sections"""] _A = clean_model_doc_toc(snake_case__) if old_modality_doc != new_modality_doc: _A = True if overwrite: _A = new_modality_doc if diff: if overwrite: _A = model_doc _A = api_doc with open(snake_case__ , """w""" , encoding="""utf-8""") as f: f.write(yaml.dump(snake_case__ , allow_unicode=snake_case__)) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""") if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') _SCREAMING_SNAKE_CASE = parser.parse_args() check_model_doc(args.fix_and_overwrite)
83
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :str , snake_case__ :PreTrainedTokenizer , snake_case__ :int , snake_case__ :Optional[int] = None , ) -> Optional[int]: _A = {} if train_file is not None: _A = [train_file] if eval_file is not None: _A = [eval_file] if test_file is not None: _A = [test_file] _A = datasets.load_dataset("""csv""" , data_files=snake_case__) _A = list(ds[list(files.keys())[0]].features.keys()) _A = features_name.pop(snake_case__) _A = list(set(ds[list(files.keys())[0]][label_name])) _A = {label: i for i, label in enumerate(snake_case__)} _A = tokenizer.model_input_names _A = {} if len(snake_case__) == 1: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( example[features_name[0]] , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""") , batched=snake_case__ , ) elif len(snake_case__) == 2: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""" , ) , batched=snake_case__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _A = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _A = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _A = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST]))) return train_ds, val_ds, test_ds, labelaid _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class a : """simple docstring""" lowerCamelCase :int = field(metadata={'''help''': '''Which column contains the label'''} ) lowerCamelCase :str = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the training file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the development file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the test file'''} ) lowerCamelCase :int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCamelCase :bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) @dataclass class a : """simple docstring""" lowerCamelCase :str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) lowerCamelCase :bool = field(default=__lowerCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) def snake_case ( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments)) _A , _A , _A = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir) and os.listdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""") # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info( F'''n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1)}, ''' F'''16-bits training: {training_args.fpaa}''') logger.info(F'''Training/evaluation parameters {training_args}''') # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _A = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _A , _A , _A , _A = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=snake_case__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _A = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(snake_case__) , labelaid=snake_case__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="""text-classification""" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _A = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(""".bin""" in model_args.model_name_or_path) , config=snake_case__ , cache_dir=model_args.cache_dir , ) def compute_metrics(snake_case__ :EvalPrediction) -> Dict: _A = np.argmax(p.predictions , axis=1) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _A = TFTrainer( model=snake_case__ , args=snake_case__ , train_dataset=snake_case__ , eval_dataset=snake_case__ , compute_metrics=snake_case__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir) # Evaluation _A = {} if training_args.do_eval: logger.info("""*** Evaluate ***""") _A = trainer.evaluate() _A = os.path.join(training_args.output_dir , """eval_results.txt""") with open(snake_case__ , """w""") as writer: logger.info("""***** Eval results *****""") for key, value in result.items(): logger.info(F''' {key} = {value}''') writer.write(F'''{key} = {value}\n''') results.update(snake_case__) return results if __name__ == "__main__": main()
83
1
def snake_case ( snake_case__ :int) -> bool: if not isinstance(snake_case__ , snake_case__): raise ValueError("""check_bouncy() accepts only integer arguments""") _A = str(snake_case__) _A = """""".join(sorted(snake_case__)) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def snake_case ( snake_case__ :float = 99) -> int: if not 0 < percent < 100: raise ValueError("""solution() only accepts values from 0 to 100""") _A = 0 _A = 1 while True: if check_bouncy(snake_case__): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(F'''{solution(99)}''')
83
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = '''speech_to_text''' lowerCamelCase :List[str] = ['''past_key_values'''] lowerCamelCase :str = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=12 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=2 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=60_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=2 , lowerCAmelCase_=(5, 5) , lowerCAmelCase_=10_24 , lowerCAmelCase_=80 , lowerCAmelCase_=1 , **lowerCAmelCase_ , ) -> Tuple: _A = vocab_size _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = encoder_layerdrop _A = decoder_layerdrop _A = use_cache _A = encoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = max_source_positions _A = max_target_positions _A = num_conv_layers _A = list(lowerCAmelCase_ ) _A = conv_channels _A = input_feat_per_channel _A = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` """ F'''but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, ''' F'''`config.num_conv_layers = {self.num_conv_layers}`.''' ) super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , )
83
1
import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format='%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s', datefmt='%Y-%m-%d %H:%M:%S', level=os.environ.get('LOGLEVEL', 'INFO').upper(), stream=sys.stdout, ) _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) _SCREAMING_SNAKE_CASE = {'facebook/bart-base': BartForConditionalGeneration} _SCREAMING_SNAKE_CASE = {'facebook/bart-base': BartTokenizer} def snake_case ( ) -> Tuple: _A = argparse.ArgumentParser(description="""Export Bart model + Beam Search to ONNX graph.""") parser.add_argument( """--validation_file""" , type=snake_case__ , default=snake_case__ , help="""A csv or a json file containing the validation data.""") parser.add_argument( """--max_length""" , type=snake_case__ , default=5 , help="""The maximum total input sequence length after tokenization.""" , ) parser.add_argument( """--num_beams""" , type=snake_case__ , default=snake_case__ , help=( """Number of beams to use for evaluation. This argument will be """ """passed to ``model.generate``, which is used during ``evaluate`` and ``predict``.""" ) , ) parser.add_argument( """--model_name_or_path""" , type=snake_case__ , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=snake_case__ , ) parser.add_argument( """--config_name""" , type=snake_case__ , default=snake_case__ , help="""Pretrained config name or path if not the same as model_name""" , ) parser.add_argument( """--device""" , type=snake_case__ , default="""cpu""" , help="""Device where the model will be run""" , ) parser.add_argument("""--output_file_path""" , type=snake_case__ , default=snake_case__ , help="""Where to store the final ONNX file.""") _A = parser.parse_args() return args def snake_case ( snake_case__ :Tuple , snake_case__ :Any="cpu") -> Dict: _A = model_dict[model_name].from_pretrained(snake_case__).to(snake_case__) _A = tokenizer_dict[model_name].from_pretrained(snake_case__) if model_name in ["facebook/bart-base"]: _A = 0 _A = None _A = 0 return huggingface_model, tokenizer def snake_case ( snake_case__ :Dict , snake_case__ :str , snake_case__ :int , snake_case__ :Dict , snake_case__ :Optional[Any]) -> Dict: model.eval() _A = None _A = torch.jit.script(BARTBeamSearchGenerator(snake_case__)) with torch.no_grad(): _A = """My friends are cool but they eat too many carbs.""" _A = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1_024 , return_tensors="""pt""").to(model.device) _A = model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , num_beams=snake_case__ , max_length=snake_case__ , early_stopping=snake_case__ , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( snake_case__ , ( inputs["""input_ids"""], inputs["""attention_mask"""], num_beams, max_length, model.config.decoder_start_token_id, ) , snake_case__ , opset_version=14 , input_names=["""input_ids""", """attention_mask""", """num_beams""", """max_length""", """decoder_start_token_id"""] , output_names=["""output_ids"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """seq"""}, """output_ids""": {0: """batch""", 1: """seq_out"""}, } , example_outputs=snake_case__ , ) logger.info("""Model exported to {}""".format(snake_case__)) _A = remove_dup_initializers(os.path.abspath(snake_case__)) logger.info("""Deduplicated and optimized model written to {}""".format(snake_case__)) _A = onnxruntime.InferenceSession(snake_case__) _A = ort_sess.run( snake_case__ , { """input_ids""": inputs["""input_ids"""].cpu().numpy(), """attention_mask""": inputs["""attention_mask"""].cpu().numpy(), """num_beams""": np.array(snake_case__), """max_length""": np.array(snake_case__), """decoder_start_token_id""": np.array(model.config.decoder_start_token_id), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1E-3 , atol=1E-3) logger.info("""Model outputs from torch and ONNX Runtime are similar.""") logger.info("""Success.""") def snake_case ( ) -> Tuple: _A = parse_args() _A = 5 _A = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.setLevel(logging.INFO) transformers.utils.logging.set_verbosity_error() _A = torch.device(args.device) _A , _A = load_model_tokenizer(args.model_name_or_path , snake_case__) if model.config.decoder_start_token_id is None: raise ValueError("""Make sure that `config.decoder_start_token_id` is correctly defined""") model.to(snake_case__) if args.max_length: _A = args.max_length if args.num_beams: _A = args.num_beams if args.output_file_path: _A = args.output_file_path else: _A = """BART.onnx""" logger.info("""Exporting model to ONNX""") export_and_validate_model(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) if __name__ == "__main__": main()
83
from __future__ import annotations from collections.abc import Callable def snake_case ( snake_case__ :Callable[[int | float], int | float] , snake_case__ :int | float , snake_case__ :int | float , snake_case__ :int = 100 , ) -> float: _A = x_start _A = fnc(snake_case__) _A = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area _A = (x_end - x_start) / steps + xa _A = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step _A = xa _A = fxa return area if __name__ == "__main__": def snake_case ( snake_case__ :Tuple) -> List[str]: return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') _SCREAMING_SNAKE_CASE = 10 while i <= 100_000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
83
1
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = ['''input_features''', '''is_longer'''] def __init__( self , lowerCAmelCase_=64 , lowerCAmelCase_=4_80_00 , lowerCAmelCase_=4_80 , lowerCAmelCase_=10 , lowerCAmelCase_=10_24 , lowerCAmelCase_=0.0 , lowerCAmelCase_=False , lowerCAmelCase_ = 0 , lowerCAmelCase_ = 1_40_00 , lowerCAmelCase_ = None , lowerCAmelCase_ = "fusion" , lowerCAmelCase_ = "repeatpad" , **lowerCAmelCase_ , ) -> Any: super().__init__( feature_size=lowerCAmelCase_ , sampling_rate=lowerCAmelCase_ , padding_value=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A = top_db _A = truncation _A = padding _A = fft_window_size _A = (fft_window_size >> 1) + 1 _A = hop_length _A = max_length_s _A = max_length_s * sampling_rate _A = sampling_rate _A = frequency_min _A = frequency_max _A = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=lowerCAmelCase_ , min_frequency=lowerCAmelCase_ , max_frequency=lowerCAmelCase_ , sampling_rate=lowerCAmelCase_ , norm=lowerCAmelCase_ , mel_scale="""htk""" , ) _A = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=lowerCAmelCase_ , min_frequency=lowerCAmelCase_ , max_frequency=lowerCAmelCase_ , sampling_rate=lowerCAmelCase_ , norm="""slaney""" , mel_scale="""slaney""" , ) def UpperCAmelCase ( self ) -> Dict[str, Any]: _A = copy.deepcopy(self.__dict__ ) _A = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> np.ndarray: _A = spectrogram( lowerCAmelCase_ , window_function(self.fft_window_size , """hann""" ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=lowerCAmelCase_ , log_mel="""dB""" , ) return log_mel_spectrogram.T def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk _A = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk _A = [0] # randomly choose index for each part _A = np.random.choice(ranges[0] ) _A = np.random.choice(ranges[1] ) _A = np.random.choice(ranges[2] ) _A = mel[idx_front : idx_front + chunk_frames, :] _A = mel[idx_middle : idx_middle + chunk_frames, :] _A = mel[idx_back : idx_back + chunk_frames, :] _A = torch.tensor(mel[None, None, :] ) _A = torch.nn.functional.interpolate( lowerCAmelCase_ , size=[chunk_frames, 64] , mode="""bilinear""" , align_corners=lowerCAmelCase_ ) _A = mel_shrink[0][0].numpy() _A = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> np.array: if waveform.shape[0] > max_length: if truncation == "rand_trunc": _A = True # random crop to max_length (for compatibility) -> this should be handled by self.pad _A = len(lowerCAmelCase_ ) - max_length _A = np.random.randint(0 , overflow + 1 ) _A = waveform[idx : idx + max_length] _A = self._np_extract_fbank_features(lowerCAmelCase_ , self.mel_filters_slaney )[None, :] elif truncation == "fusion": _A = self._np_extract_fbank_features(lowerCAmelCase_ , self.mel_filters ) _A = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed _A = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. _A = np.stack([mel, mel, mel, mel] , axis=0 ) _A = False else: _A = self._random_mel_fusion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = True else: raise NotImplementedError(F'''data_truncating {truncation} not implemented''' ) else: _A = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": _A = int(max_length / len(lowerCAmelCase_ ) ) _A = np.stack(np.tile(lowerCAmelCase_ , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": _A = int(max_length / len(lowerCAmelCase_ ) ) _A = np.stack(np.tile(lowerCAmelCase_ , lowerCAmelCase_ ) ) _A = np.pad(lowerCAmelCase_ , (0, max_length - waveform.shape[0]) , mode="""constant""" , constant_values=0 ) if truncation == "fusion": _A = self._np_extract_fbank_features(lowerCAmelCase_ , self.mel_filters ) _A = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: _A = self._np_extract_fbank_features(lowerCAmelCase_ , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , **lowerCAmelCase_ , ) -> BatchFeature: _A = truncation if truncation is not None else self.truncation _A = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' F''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' F''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) _A = isinstance(lowerCAmelCase_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) _A = is_batched_numpy or ( isinstance(lowerCAmelCase_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _A = [np.asarray(lowerCAmelCase_ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase_ , np.ndarray ): _A = np.asarray(lowerCAmelCase_ , dtype=np.floataa ) elif isinstance(lowerCAmelCase_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _A = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _A = [np.asarray(lowerCAmelCase_ )] # convert to mel spectrogram, truncate and pad if needed. _A = [ self._get_input_mel(lowerCAmelCase_ , max_length if max_length else self.nb_max_samples , lowerCAmelCase_ , lowerCAmelCase_ ) for waveform in raw_speech ] _A = [] _A = [] for mel, longer in padded_inputs: input_mel.append(lowerCAmelCase_ ) is_longer.append(lowerCAmelCase_ ) if truncation == "fusion" and sum(lowerCAmelCase_ ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer _A = np.random.randint(0 , len(lowerCAmelCase_ ) ) _A = True if isinstance(input_mel[0] , lowerCAmelCase_ ): _A = [np.asarray(lowerCAmelCase_ , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool _A = [[longer] for longer in is_longer] _A = {"""input_features""": input_mel, """is_longer""": is_longer} _A = BatchFeature(lowerCAmelCase_ ) if return_tensors is not None: _A = input_features.convert_to_tensors(lowerCAmelCase_ ) return input_features
83
import numpy as np import qiskit def snake_case ( snake_case__ :int = 8 , snake_case__ :int | None = None) -> str: _A = np.random.default_rng(seed=snake_case__) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _A = 6 * key_len # Measurement basis for Alice's qubits. _A = rng.integers(2 , size=snake_case__) # The set of states Alice will prepare. _A = rng.integers(2 , size=snake_case__) # Measurement basis for Bob's qubits. _A = rng.integers(2 , size=snake_case__) # Quantum Circuit to simulate BB84 _A = qiskit.QuantumCircuit(snake_case__ , name="""BB84""") # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case__): if alice_state[index] == 1: bbaa_circ.x(snake_case__) if alice_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case__): if bob_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _A = qiskit.Aer.get_backend("""aer_simulator""") # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _A = qiskit.execute(snake_case__ , snake_case__ , shots=1 , seed_simulator=snake_case__) # Returns the result of measurement. _A = job.result().get_counts(snake_case__).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _A = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case__ , snake_case__ , snake_case__) if alice_basis_bit == bob_basis_bit ]) # Get final key. Pad with 0 if too short, otherwise truncate. _A = gen_key[:key_len] if len(snake_case__) >= key_len else gen_key.ljust(snake_case__ , """0""") return key if __name__ == "__main__": print(F'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
83
1
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _SCREAMING_SNAKE_CASE = datasets.logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _SCREAMING_SNAKE_CASE = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _SCREAMING_SNAKE_CASE = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def snake_case ( snake_case__ :Optional[Any] , snake_case__ :str , snake_case__ :List[str]=False , snake_case__ :Dict=False , snake_case__ :Any=True , snake_case__ :List[str]=False , snake_case__ :Optional[Any]="dummy_doc") -> List[Any]: _A = {doc: key_lines} _A = {doc: sys_lines} _A = {} _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A , _A = reader.get_doc_mentions(snake_case__ , key_doc_lines[doc] , snake_case__) key_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) _A , _A = reader.get_doc_mentions(snake_case__ , sys_doc_lines[doc] , snake_case__) sys_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) if remove_nested: _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F'''annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}''') logger.info( """Number of resulting singleton clusters in the key """ F'''annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}''') if not keep_singletons: logger.info( F'''{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ''' """files, respectively""") return doc_coref_infos def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Dict , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Tuple) -> int: _A = get_coref_infos(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) _A = {} _A = 0 _A = 0 for name, metric in metrics: _A , _A , _A = evaluator.evaluate_documents(snake_case__ , snake_case__ , beta=1) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'''{name}/recall''': recall, F'''{name}/precision''': precision, F'''{name}/f1''': fa}) logger.info( name.ljust(10) , F'''Recall: {recall * 100:.2f}''' , F''' Precision: {precision * 100:.2f}''' , F''' F1: {fa * 100:.2f}''' , ) if conll_subparts_num == 3: _A = (conll / 3) * 100 logger.info(F'''CoNLL score: {conll:.2f}''') output_scores.update({"""conll_score""": conll}) return output_scores def snake_case ( snake_case__ :Union[str, Any]) -> List[Any]: _A = False for line in key_lines: if not line.startswith("""#"""): if len(line.split()) > 6: _A = line.split()[5] if not parse_col == "-": _A = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Union[str, Any]: _A = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _A = util.check_gold_parse_annotation(lowerCAmelCase_ ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _A = evaluate( key_lines=lowerCAmelCase_ , sys_lines=lowerCAmelCase_ , metrics=lowerCAmelCase_ , NP_only=lowerCAmelCase_ , remove_nested=lowerCAmelCase_ , keep_singletons=lowerCAmelCase_ , min_span=lowerCAmelCase_ , ) return score
83
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def snake_case ( snake_case__ :int) -> Optional[int]: return EnvironmentCommand() def snake_case ( snake_case__ :Tuple) -> List[str]: return EnvironmentCommand(args.accelerate_config_file) class a ( __lowerCAmelCase ): """simple docstring""" @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: _A = parser.add_parser("""env""" ) download_parser.set_defaults(func=lowerCAmelCase_ ) download_parser.add_argument( """--accelerate-config_file""" , default=lowerCAmelCase_ , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=lowerCAmelCase_ ) def __init__( self , lowerCAmelCase_ , *lowerCAmelCase_ ) -> None: _A = accelerate_config_file def UpperCAmelCase ( self ) -> Dict: _A = """not installed""" if is_safetensors_available(): import safetensors _A = safetensors.__version__ elif importlib.util.find_spec("""safetensors""" ) is not None: import safetensors _A = F'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' _A = """not installed""" _A = _A = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file _A = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(lowerCAmelCase_ ): _A = load_config_from_file(self._accelerate_config_file ).to_dict() _A = ( """\n""".join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else F'''\t{accelerate_config}''' ) _A = """not installed""" _A = """NA""" if is_torch_available(): import torch _A = torch.__version__ _A = torch.cuda.is_available() _A = """not installed""" _A = """NA""" if is_tf_available(): import tensorflow as tf _A = tf.__version__ try: # deprecated in v2.1 _A = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool _A = bool(tf.config.list_physical_devices("""GPU""" ) ) _A = """not installed""" _A = """not installed""" _A = """not installed""" _A = """NA""" if is_flax_available(): import flax import jax import jaxlib _A = flax.__version__ _A = jax.__version__ _A = jaxlib.__version__ _A = jax.lib.xla_bridge.get_backend().platform _A = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": F'''{safetensors_version}''', """Accelerate version""": F'''{accelerate_version}''', """Accelerate config""": F'''{accelerate_config_str}''', """PyTorch version (GPU?)""": F'''{pt_version} ({pt_cuda_available})''', """Tensorflow version (GPU?)""": F'''{tf_version} ({tf_cuda_available})''', """Flax version (CPU?/GPU?/TPU?)""": F'''{flax_version} ({jax_backend})''', """Jax version""": F'''{jax_version}''', """JaxLib version""": F'''{jaxlib_version}''', """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""" ) print(self.format_dict(lowerCAmelCase_ ) ) return info @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: return "\n".join([F'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
83
1
from collections import defaultdict def snake_case ( snake_case__ :str , snake_case__ :str) -> bool: _A = first_str.lower().strip() _A = second_str.lower().strip() # Remove whitespace _A = first_str.replace(""" """ , """""") _A = second_str.replace(""" """ , """""") # Strings of different lengths are not anagrams if len(snake_case__) != len(snake_case__): return False # Default values for count should be 0 _A = defaultdict(snake_case__) # For each character in input strings, # increment count in the corresponding for i in range(len(snake_case__)): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values()) if __name__ == "__main__": from doctest import testmod testmod() _SCREAMING_SNAKE_CASE = input('Enter the first string ').strip() _SCREAMING_SNAKE_CASE = input('Enter the second string ').strip() _SCREAMING_SNAKE_CASE = check_anagrams(input_a, input_b) print(F'''{input_a} and {input_b} are {"" if status else "not "}anagrams.''')
83
import colorsys from PIL import Image # type: ignore def snake_case ( snake_case__ :float , snake_case__ :float , snake_case__ :int) -> float: _A = x _A = y for step in range(snake_case__): # noqa: B007 _A = a * a - b * b + x _A = 2 * a * b + y _A = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255) for i in colorsys.hsv_to_rgb(snake_case__ , 1 , 1)) def snake_case ( snake_case__ :int = 800 , snake_case__ :int = 600 , snake_case__ :float = -0.6 , snake_case__ :float = 0 , snake_case__ :float = 3.2 , snake_case__ :int = 50 , snake_case__ :bool = True , ) -> Image.Image: _A = Image.new("""RGB""" , (image_width, image_height)) _A = img.load() # loop through the image-coordinates for image_x in range(snake_case__): for image_y in range(snake_case__): # determine the figure-coordinates based on the image-coordinates _A = figure_width / image_width * image_height _A = figure_center_x + (image_x / image_width - 0.5) * figure_width _A = figure_center_y + (image_y / image_height - 0.5) * figure_height _A = get_distance(snake_case__ , snake_case__ , snake_case__) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _A = get_color_coded_rgb(snake_case__) else: _A = get_black_and_white_rgb(snake_case__) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _SCREAMING_SNAKE_CASE = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
1
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Optional[int]: _A = tempfile.mkdtemp() _A = 5 # Realm tok _A = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """test""", """question""", """this""", """is""", """the""", """first""", """second""", """third""", """fourth""", """fifth""", """record""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] _A = os.path.join(self.tmpdirname , """realm_tokenizer""" ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) _A = os.path.join(lowerCAmelCase_ , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) _A = os.path.join(self.tmpdirname , """realm_block_records""" ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> RealmTokenizer: return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , """realm_tokenizer""" ) ) def UpperCAmelCase ( self ) -> Any: shutil.rmtree(self.tmpdirname ) def UpperCAmelCase ( self ) -> Any: _A = RealmConfig(num_block_records=self.num_block_records ) return config def UpperCAmelCase ( self ) -> List[Any]: _A = Dataset.from_dict( { """id""": ["""0""", """1"""], """question""": ["""foo""", """bar"""], """answers""": [["""Foo""", """Bar"""], ["""Bar"""]], } ) return dataset def UpperCAmelCase ( self ) -> List[str]: _A = np.array( [ B"""This is the first record""", B"""This is the second record""", B"""This is the third record""", B"""This is the fourth record""", B"""This is the fifth record""", B"""This is a longer longer longer record""", ] , dtype=lowerCAmelCase_ , ) return block_records def UpperCAmelCase ( self ) -> str: _A = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def UpperCAmelCase ( self ) -> Optional[Any]: _A = self.get_config() _A = self.get_dummy_retriever() _A = retriever.tokenizer _A = np.array([0, 3] , dtype="""long""" ) _A = tokenizer(["""Test question"""] ).input_ids _A = tokenizer( ["""the fourth"""] , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ).input_ids _A = config.reader_seq_len _A , _A , _A , _A = retriever( lowerCAmelCase_ , lowerCAmelCase_ , answer_ids=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors="""np""" ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ["""[CLS]""", """test""", """question""", """[SEP]""", """this""", """is""", """the""", """first""", """record""", """[SEP]"""] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ["""[CLS]""", """test""", """question""", """[SEP]""", """this""", """is""", """the""", """fourth""", """record""", """[SEP]"""] , ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = self.get_config() _A = self.get_dummy_retriever() _A = retriever.tokenizer _A = np.array([0, 3, 5] , dtype="""long""" ) _A = tokenizer(["""Test question"""] ).input_ids _A = tokenizer( ["""the fourth""", """longer longer"""] , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ).input_ids _A = config.reader_seq_len _A , _A , _A , _A = retriever( lowerCAmelCase_ , lowerCAmelCase_ , answer_ids=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors="""np""" ) self.assertEqual([False, True, True] , lowerCAmelCase_ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , lowerCAmelCase_ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , """realm_block_records""" ) ) # Test local path _A = retriever.from_pretrained(os.path.join(self.tmpdirname , """realm_block_records""" ) ) self.assertEqual(retriever.block_records[0] , B"""This is the first record""" ) # Test mocked remote path with patch("""transformers.models.realm.retrieval_realm.hf_hub_download""" ) as mock_hf_hub_download: _A = os.path.join( os.path.join(self.tmpdirname , """realm_block_records""" ) , _REALM_BLOCK_RECORDS_FILENAME ) _A = RealmRetriever.from_pretrained("""google/realm-cc-news-pretrained-openqa""" ) self.assertEqual(retriever.block_records[0] , B"""This is the first record""" )
83
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _SCREAMING_SNAKE_CASE = datasets.logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _SCREAMING_SNAKE_CASE = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _SCREAMING_SNAKE_CASE = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def snake_case ( snake_case__ :Optional[Any] , snake_case__ :str , snake_case__ :List[str]=False , snake_case__ :Dict=False , snake_case__ :Any=True , snake_case__ :List[str]=False , snake_case__ :Optional[Any]="dummy_doc") -> List[Any]: _A = {doc: key_lines} _A = {doc: sys_lines} _A = {} _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A , _A = reader.get_doc_mentions(snake_case__ , key_doc_lines[doc] , snake_case__) key_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) _A , _A = reader.get_doc_mentions(snake_case__ , sys_doc_lines[doc] , snake_case__) sys_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) if remove_nested: _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F'''annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}''') logger.info( """Number of resulting singleton clusters in the key """ F'''annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}''') if not keep_singletons: logger.info( F'''{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ''' """files, respectively""") return doc_coref_infos def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Dict , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Tuple) -> int: _A = get_coref_infos(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) _A = {} _A = 0 _A = 0 for name, metric in metrics: _A , _A , _A = evaluator.evaluate_documents(snake_case__ , snake_case__ , beta=1) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'''{name}/recall''': recall, F'''{name}/precision''': precision, F'''{name}/f1''': fa}) logger.info( name.ljust(10) , F'''Recall: {recall * 100:.2f}''' , F''' Precision: {precision * 100:.2f}''' , F''' F1: {fa * 100:.2f}''' , ) if conll_subparts_num == 3: _A = (conll / 3) * 100 logger.info(F'''CoNLL score: {conll:.2f}''') output_scores.update({"""conll_score""": conll}) return output_scores def snake_case ( snake_case__ :Union[str, Any]) -> List[Any]: _A = False for line in key_lines: if not line.startswith("""#"""): if len(line.split()) > 6: _A = line.split()[5] if not parse_col == "-": _A = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Union[str, Any]: _A = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _A = util.check_gold_parse_annotation(lowerCAmelCase_ ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _A = evaluate( key_lines=lowerCAmelCase_ , sys_lines=lowerCAmelCase_ , metrics=lowerCAmelCase_ , NP_only=lowerCAmelCase_ , remove_nested=lowerCAmelCase_ , keep_singletons=lowerCAmelCase_ , min_span=lowerCAmelCase_ , ) return score
83
1
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=2 , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=2 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=36 , lowerCAmelCase_=2 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=6 , lowerCAmelCase_=6 , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=None , lowerCAmelCase_=10_00 , ) -> Any: _A = parent _A = batch_size _A = num_channels _A = image_size _A = patch_size _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = coordinate_size _A = shape_size _A = num_labels _A = num_choices _A = scope _A = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _A = text_seq_length _A = (image_size // patch_size) ** 2 + 1 _A = self.text_seq_length + self.image_seq_length def UpperCAmelCase ( self ) -> Dict: _A = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _A = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) _A = bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _A = bbox[i, j, 3] _A = bbox[i, j, 1] _A = tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: _A = bbox[i, j, 2] _A = bbox[i, j, 0] _A = tmp_coordinate _A = tf.constant(lowerCAmelCase_ ) _A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _A = None if self.use_input_mask: _A = random_attention_mask([self.batch_size, self.text_seq_length] ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _A = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: _A = TFLayoutLMvaModel(config=lowerCAmelCase_ ) # text + image _A = model(lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , training=lowerCAmelCase_ ) _A = model( lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , training=lowerCAmelCase_ , ) _A = model(lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , training=lowerCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _A = model(lowerCAmelCase_ , training=lowerCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _A = model({"""pixel_values""": pixel_values} , training=lowerCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = self.num_labels _A = TFLayoutLMvaForSequenceClassification(config=lowerCAmelCase_ ) _A = model( lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , training=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Union[str, Any]: _A = self.num_labels _A = TFLayoutLMvaForTokenClassification(config=lowerCAmelCase_ ) _A = model( lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , training=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _A = 2 _A = TFLayoutLMvaForQuestionAnswering(config=lowerCAmelCase_ ) _A = model( lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , training=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self ) -> Optional[Any]: _A = self.prepare_config_and_inputs() ((_A) , (_A) , (_A) , (_A) , (_A) , (_A) , (_A) , (_A)) = config_and_inputs _A = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_tf class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :str = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) lowerCamelCase :Any = ( {'''document-question-answering''': TFLayoutLMvaForQuestionAnswering, '''feature-extraction''': TFLayoutLMvaModel} if is_tf_available() else {} ) lowerCamelCase :Optional[Any] = False lowerCamelCase :List[Any] = False lowerCamelCase :int = False def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: return True def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False ) -> dict: _A = copy.deepcopy(lowerCAmelCase_ ) if model_class in get_values(lowerCAmelCase_ ): _A = { k: tf.tile(tf.expand_dims(lowerCAmelCase_ , 1 ) , (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(lowerCAmelCase_ , tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(lowerCAmelCase_ ): _A = tf.ones(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(lowerCAmelCase_ ): _A = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) _A = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(lowerCAmelCase_ ): _A = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(lowerCAmelCase_ ): _A = tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=tf.intaa ) return inputs_dict def UpperCAmelCase ( self ) -> List[str]: _A = TFLayoutLMvaModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> List[Any]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> Tuple: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(lowerCAmelCase_ ) if getattr(lowerCAmelCase_ , """hf_compute_loss""" , lowerCAmelCase_ ): # The number of elements in the loss should be the same as the number of elements in the label _A = self._prepare_for_class(inputs_dict.copy() , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) _A = prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys() , reverse=lowerCAmelCase_ )[0] ] _A = added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs _A = self._prepare_for_class(inputs_dict.copy() , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) _A = prepared_for_class.pop("""input_ids""" ) _A = model(lowerCAmelCase_ , **lowerCAmelCase_ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions _A = self._prepare_for_class(inputs_dict.copy() , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) _A = prepared_for_class.pop("""input_ids""" ) if "labels" in prepared_for_class: _A = prepared_for_class["""labels"""].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: _A = -1_00 _A = tf.convert_to_tensor(lowerCAmelCase_ ) _A = model(lowerCAmelCase_ , **lowerCAmelCase_ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict _A = self._prepare_for_class(inputs_dict.copy() , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) _A = model(lowerCAmelCase_ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple _A = self._prepare_for_class(inputs_dict.copy() , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) # Get keys that were added with the _prepare_for_class function _A = prepared_for_class.keys() - inputs_dict.keys() _A = inspect.signature(model.call ).parameters _A = list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple _A = {0: """input_ids"""} for label_key in label_keys: _A = signature_names.index(lowerCAmelCase_ ) _A = label_key _A = sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple _A = [] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: _A = prepared_for_class[value] _A = tuple(lowerCAmelCase_ ) # Send to model _A = model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def UpperCAmelCase ( self ) -> Dict: ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[Any]: ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _A = type self.model_tester.create_and_check_model(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> List[Any]: for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = TFLayoutLMvaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def snake_case ( ) -> str: _A = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""") return image @require_tf class a ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase ( self ) -> List[Any]: return LayoutLMvaImageProcessor(apply_ocr=lowerCAmelCase_ ) if is_vision_available() else None @slow def UpperCAmelCase ( self ) -> Tuple: _A = TFLayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ) _A = self.default_image_processor _A = prepare_img() _A = image_processor(images=lowerCAmelCase_ , return_tensors="""tf""" ).pixel_values _A = tf.constant([[1, 2]] ) _A = tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ) , axis=0 ) # forward pass _A = model(input_ids=lowerCAmelCase_ , bbox=lowerCAmelCase_ , pixel_values=lowerCAmelCase_ , training=lowerCAmelCase_ ) # verify the logits _A = (1, 1_99, 7_68) self.assertEqual(outputs.last_hidden_state.shape , lowerCAmelCase_ ) _A = tf.constant( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCAmelCase_ , atol=1E-4 ) )
83
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } _SCREAMING_SNAKE_CASE = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } _SCREAMING_SNAKE_CASE = {'facebook/blenderbot_small-90M': 512} def snake_case ( snake_case__ :Tuple) -> str: _A = set() _A = word[0] for char in word[1:]: pairs.add((prev_char, char)) _A = char _A = set(snake_case__) return pairs class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :List[Any] = VOCAB_FILES_NAMES lowerCamelCase :Tuple = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase :List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase :int = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="__start__" , lowerCAmelCase_="__end__" , lowerCAmelCase_="__unk__" , lowerCAmelCase_="__null__" , **lowerCAmelCase_ , ) -> int: super().__init__(unk_token=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , **lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) _A = {v: k for k, v in self.encoder.items()} with open(lowerCAmelCase_ , encoding="""utf-8""" ) as merges_handle: _A = merges_handle.read().split("""\n""" )[1:-1] _A = [tuple(merge.split() ) for merge in merges] _A = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_ ) ) ) ) _A = {} @property def UpperCAmelCase ( self ) -> int: return len(self.encoder ) def UpperCAmelCase ( self ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: if token in self.cache: return self.cache[token] _A = re.sub("""([.,!?()])""" , r""" \1""" , lowerCAmelCase_ ) _A = re.sub("""(')""" , r""" \1 """ , lowerCAmelCase_ ) _A = re.sub(r"""\s{2,}""" , """ """ , lowerCAmelCase_ ) if "\n" in token: _A = token.replace("""\n""" , """ __newln__""" ) _A = token.split(""" """ ) _A = [] for token in tokens: if not len(lowerCAmelCase_ ): continue _A = token.lower() _A = tuple(lowerCAmelCase_ ) _A = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) _A = get_pairs(lowerCAmelCase_ ) if not pairs: words.append(lowerCAmelCase_ ) continue while True: _A = min(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : self.bpe_ranks.get(lowerCAmelCase_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break _A , _A = bigram _A = [] _A = 0 while i < len(lowerCAmelCase_ ): try: _A = word.index(lowerCAmelCase_ , lowerCAmelCase_ ) new_word.extend(word[i:j] ) _A = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A = tuple(lowerCAmelCase_ ) _A = new_word if len(lowerCAmelCase_ ) == 1: break else: _A = get_pairs(lowerCAmelCase_ ) _A = """@@ """.join(lowerCAmelCase_ ) _A = word[:-4] _A = word words.append(lowerCAmelCase_ ) return " ".join(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[str]: _A = [] _A = re.findall(r"""\S+\n?""" , lowerCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(lowerCAmelCase_ ).split(""" """ ) ) ) return split_tokens def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: _A = token.lower() return self.encoder.get(lowerCAmelCase_ , self.encoder.get(self.unk_token ) ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: return self.decoder.get(lowerCAmelCase_ , self.unk_token ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = """ """.join(lowerCAmelCase_ ).replace("""@@ """ , """""" ).strip() return out_string def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase_ , ensure_ascii=lowerCAmelCase_ ) + """\n""" ) _A = 0 with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase_ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) _A = token_index writer.write(""" """.join(lowerCAmelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file
83
1
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = TypeVar('DatasetType', Dataset, IterableDataset) def snake_case ( snake_case__ :List[DatasetType] , snake_case__ :Optional[List[float]] = None , snake_case__ :Optional[int] = None , snake_case__ :Optional[DatasetInfo] = None , snake_case__ :Optional[NamedSplit] = None , snake_case__ :Literal["first_exhausted", "all_exhausted"] = "first_exhausted" , ) -> DatasetType: from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError("""Unable to interleave an empty list of datasets.""") for i, dataset in enumerate(snake_case__): if not isinstance(snake_case__ , (Dataset, IterableDataset)): if isinstance(snake_case__ , (DatasetDict, IterableDatasetDict)): if not dataset: raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ''' """is an empty dataset dictionary.""") raise ValueError( F'''Dataset at position {i} has at least one split: {list(snake_case__)}\n''' F'''Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(snake_case__))}\']''') raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(snake_case__).__name__}.''') if i == 0: _A , _A = ( (Dataset, IterableDataset) if isinstance(snake_case__ , snake_case__) else (IterableDataset, Dataset) ) elif not isinstance(snake_case__ , snake_case__): raise ValueError( F'''Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.''') if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F'''{stopping_strategy} is not supported. Please enter a valid stopping_strategy.''') if dataset_type is Dataset: return _interleave_map_style_datasets( snake_case__ , snake_case__ , snake_case__ , info=snake_case__ , split=snake_case__ , stopping_strategy=snake_case__) else: return _interleave_iterable_datasets( snake_case__ , snake_case__ , snake_case__ , info=snake_case__ , split=snake_case__ , stopping_strategy=snake_case__) def snake_case ( snake_case__ :List[DatasetType] , snake_case__ :Optional[DatasetInfo] = None , snake_case__ :Optional[NamedSplit] = None , snake_case__ :int = 0 , ) -> DatasetType: if not dsets: raise ValueError("""Unable to concatenate an empty list of datasets.""") for i, dataset in enumerate(snake_case__): if not isinstance(snake_case__ , (Dataset, IterableDataset)): if isinstance(snake_case__ , (DatasetDict, IterableDatasetDict)): if not dataset: raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ''' """is an empty dataset dictionary.""") raise ValueError( F'''Dataset at position {i} has at least one split: {list(snake_case__)}\n''' F'''Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(snake_case__))}\']''') raise ValueError( F'''Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(snake_case__).__name__}.''') if i == 0: _A , _A = ( (Dataset, IterableDataset) if isinstance(snake_case__ , snake_case__) else (IterableDataset, Dataset) ) elif not isinstance(snake_case__ , snake_case__): raise ValueError( F'''Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.''') if dataset_type is Dataset: return _concatenate_map_style_datasets(snake_case__ , info=snake_case__ , split=snake_case__ , axis=snake_case__) else: return _concatenate_iterable_datasets(snake_case__ , info=snake_case__ , split=snake_case__ , axis=snake_case__)
83
_SCREAMING_SNAKE_CASE = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on _SCREAMING_SNAKE_CASE = {value: key for key, value in MORSE_CODE_DICT.items()} def snake_case ( snake_case__ :str) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper()) def snake_case ( snake_case__ :str) -> str: return "".join(REVERSE_DICT[char] for char in message.split()) def snake_case ( ) -> None: _A = """Morse code here!""" print(snake_case__) _A = encrypt(snake_case__) print(snake_case__) _A = decrypt(snake_case__) print(snake_case__) if __name__ == "__main__": main()
83
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :Any = StableDiffusionXLImgaImgPipeline lowerCamelCase :List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} lowerCamelCase :List[str] = PipelineTesterMixin.required_optional_params - {'''latents'''} lowerCamelCase :Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCamelCase :Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase :Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase ( self ) -> Any: torch.manual_seed(0 ) _A = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , attention_head_dim=(2, 4) , use_linear_projection=lowerCAmelCase_ , addition_embed_type="""text_time""" , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) _A = EulerDiscreteScheduler( beta_start=0.0_0085 , beta_end=0.012 , steps_offset=1 , beta_schedule="""scaled_linear""" , timestep_spacing="""leading""" , ) torch.manual_seed(0 ) _A = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=32 , ) _A = CLIPTextModel(lowerCAmelCase_ ) _A = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" , local_files_only=lowerCAmelCase_ ) _A = CLIPTextModelWithProjection(lowerCAmelCase_ ) _A = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" , local_files_only=lowerCAmelCase_ ) _A = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """text_encoder_2""": text_encoder_a, """tokenizer_2""": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=0 ) -> Optional[Any]: _A = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) _A = image / 2 + 0.5 if str(lowerCAmelCase_ ).startswith("""mps""" ): _A = torch.manual_seed(lowerCAmelCase_ ) else: _A = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _A = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 5.0, """output_type""": """numpy""", """strength""": 0.75, } return inputs def UpperCAmelCase ( self ) -> int: _A = """cpu""" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase_ ) _A = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = self.get_dummy_inputs(lowerCAmelCase_ ) _A = sd_pipe(**lowerCAmelCase_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _A = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase ( self ) -> str: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCAmelCase ( self ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCAmelCase ( self ) -> Optional[Any]: pass def UpperCAmelCase ( self ) -> int: _A = self.get_dummy_components() _A = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase_ ) _A = sd_pipe.to(lowerCAmelCase_ ) _A = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) # forward without prompt embeds _A = self.get_dummy_inputs(lowerCAmelCase_ ) _A = 3 * ["""this is a negative prompt"""] _A = negative_prompt _A = 3 * [inputs["""prompt"""]] _A = sd_pipe(**lowerCAmelCase_ ) _A = output.images[0, -3:, -3:, -1] # forward with prompt embeds _A = self.get_dummy_inputs(lowerCAmelCase_ ) _A = 3 * ["""this is a negative prompt"""] _A = 3 * [inputs.pop("""prompt""" )] ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = sd_pipe.encode_prompt(lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) _A = sd_pipe( **lowerCAmelCase_ , prompt_embeds=lowerCAmelCase_ , negative_prompt_embeds=lowerCAmelCase_ , pooled_prompt_embeds=lowerCAmelCase_ , negative_pooled_prompt_embeds=lowerCAmelCase_ , ) _A = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_="cpu" , lowerCAmelCase_=torch.floataa , lowerCAmelCase_=0 ) -> Tuple: _A = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _A = np.random.RandomState(lowerCAmelCase_ ).standard_normal((1, 4, 64, 64) ) _A = torch.from_numpy(lowerCAmelCase_ ).to(device=lowerCAmelCase_ , dtype=lowerCAmelCase_ ) _A = { """prompt""": """a photograph of an astronaut riding a horse""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def UpperCAmelCase ( self ) -> int: _A = DiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-base""" ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = self.get_inputs(lowerCAmelCase_ ) _A = pipe(**lowerCAmelCase_ ).images _A = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) _A = np.array([0.4_9493, 0.4_7896, 0.4_0798, 0.5_4214, 0.5_3212, 0.4_8202, 0.4_7656, 0.4_6329, 0.4_8506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
83
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _SCREAMING_SNAKE_CASE = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :int = (KDPMaDiscreteScheduler,) lowerCamelCase :str = 10 def UpperCAmelCase ( self , **lowerCAmelCase_ ) -> str: _A = { """num_train_timesteps""": 11_00, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**lowerCAmelCase_ ) return config def UpperCAmelCase ( self ) -> str: for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[Any]: for beta_start, beta_end in zip([0.0_0001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=lowerCAmelCase_ , beta_end=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Union[str, Any]: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: _A = self.scheduler_classes[0] _A = self.get_scheduler_config(prediction_type="""v_prediction""" ) _A = scheduler_class(**lowerCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) _A = self.dummy_model() _A = self.dummy_sample_deter * scheduler.init_noise_sigma _A = sample.to(lowerCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): _A = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_ ) _A = model(lowerCAmelCase_ , lowerCAmelCase_ ) _A = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = output.prev_sample _A = torch.sum(torch.abs(lowerCAmelCase_ ) ) _A = torch.mean(torch.abs(lowerCAmelCase_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693_4286_5017_0972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0002 ) < 1E-3 def UpperCAmelCase ( self ) -> str: if torch_device == "mps": return _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**lowerCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) _A = self.dummy_model() _A = self.dummy_sample_deter * scheduler.init_noise_sigma _A = sample.to(lowerCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): _A = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_ ) _A = model(lowerCAmelCase_ , lowerCAmelCase_ ) _A = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = output.prev_sample _A = torch.sum(torch.abs(lowerCAmelCase_ ) ) _A = torch.mean(torch.abs(lowerCAmelCase_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 def UpperCAmelCase ( self ) -> List[Any]: if torch_device == "mps": return _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**lowerCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=lowerCAmelCase_ ) _A = self.dummy_model() _A = self.dummy_sample_deter.to(lowerCAmelCase_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: _A = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_ ) _A = model(lowerCAmelCase_ , lowerCAmelCase_ ) _A = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = output.prev_sample _A = torch.sum(torch.abs(lowerCAmelCase_ ) ) _A = torch.mean(torch.abs(lowerCAmelCase_ ) ) if str(lowerCAmelCase_ ).startswith("""cpu""" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1E-2 assert abs(result_mean.item() - 0.0266 ) < 1E-3
83
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''philschmid/bart-large-cnn-samsum''' lowerCamelCase :Tuple = ( '''This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ''' '''and returns a summary of the text.''' ) lowerCamelCase :List[Any] = '''summarizer''' lowerCamelCase :List[str] = AutoTokenizer lowerCamelCase :Dict = AutoModelForSeqaSeqLM lowerCamelCase :int = ['''text'''] lowerCamelCase :List[Any] = ['''text'''] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: return self.pre_processor(lowerCAmelCase_ , return_tensors="""pt""" , truncation=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: return self.model.generate(**lowerCAmelCase_ )[0] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: return self.pre_processor.decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ )
83
1
from __future__ import annotations from PIL import Image # Define glider example _SCREAMING_SNAKE_CASE = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example _SCREAMING_SNAKE_CASE = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def snake_case ( snake_case__ :list[list[int]]) -> list[list[int]]: _A = [] for i in range(len(snake_case__)): _A = [] for j in range(len(cells[i])): # Get the number of live neighbours _A = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i]) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i]) - 1: neighbour_count += cells[i][j + 1] if i < len(snake_case__) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(snake_case__) - 1: neighbour_count += cells[i + 1][j] if i < len(snake_case__) - 1 and j < len(cells[i]) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. _A = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1) else: next_generation_row.append(0) next_generation.append(snake_case__) return next_generation def snake_case ( snake_case__ :list[list[int]] , snake_case__ :int) -> list[Image.Image]: _A = [] for _ in range(snake_case__): # Create output image _A = Image.new("""RGB""" , (len(cells[0]), len(snake_case__))) _A = img.load() # Save cells to image for x in range(len(snake_case__)): for y in range(len(cells[0])): _A = 255 - cells[y][x] * 255 _A = (colour, colour, colour) # Save image images.append(snake_case__) _A = new_generation(snake_case__) return images if __name__ == "__main__": _SCREAMING_SNAKE_CASE = generate_images(GLIDER, 16) images[0].save('out.gif', save_all=True, append_images=images[1:])
83
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _SCREAMING_SNAKE_CASE = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def snake_case ( snake_case__ :Union[str, Any]) -> Dict: _A = torch.load(snake_case__ , map_location="""cpu""") return sd def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any] , snake_case__ :int=rename_keys_prefix) -> Optional[Any]: _A = OrderedDict() _A = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _A = key for name_pair in rename_keys_prefix: _A = new_key.replace(name_pair[0] , name_pair[1]) _A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _A = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple) -> int: assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _A = """pretraining""" if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "nlvr" in checkpoint_path: _A = {"""visual_embedding_dim""": 1_024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} _A = """multichoice""" elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} _A = """vqa_advanced""" elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048, """num_labels""": 3_129} _A = """vqa""" elif "nlvr" in checkpoint_path: _A = { """visual_embedding_dim""": 1_024, """num_labels""": 2, } _A = """nlvr""" _A = VisualBertConfig(**snake_case__) # Load State Dict _A = load_state_dict(snake_case__) _A = get_new_dict(snake_case__ , snake_case__) if model_type == "pretraining": _A = VisualBertForPreTraining(snake_case__) elif model_type == "vqa": _A = VisualBertForQuestionAnswering(snake_case__) elif model_type == "nlvr": _A = VisualBertForVisualReasoning(snake_case__) elif model_type == "multichoice": _A = VisualBertForMultipleChoice(snake_case__) model.load_state_dict(snake_case__) # Save Checkpoints Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
83
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = True , lowerCAmelCase_ = None , lowerCAmelCase_ = False , lowerCAmelCase_ = None , lowerCAmelCase_ = True , lowerCAmelCase_ = "arrow" , **lowerCAmelCase_ , ) -> Dict: super().__init__( split=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ , streaming=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A = load_from_cache_file _A = file_format _A = Spark( df=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , working_dir=lowerCAmelCase_ , **lowerCAmelCase_ , ) def UpperCAmelCase ( self ) -> str: if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) _A = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowerCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
83
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def UpperCAmelCase ( self ) -> Optional[int]: _A = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCAmelCase_ ): self.assertDictEqual(lowerCAmelCase_ , example_records[i] ) def UpperCAmelCase ( self ) -> str: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) _A = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def UpperCAmelCase ( self ) -> Any: # checks what happens with missing columns _A = [{"""col_1""": 1}, {"""col_2""": """x"""}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def UpperCAmelCase ( self ) -> Tuple: # checks if the type can be inferred from the second record _A = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def UpperCAmelCase ( self ) -> Any: _A = Dataset.from_list([] ) self.assertEqual(len(lowerCAmelCase_ ) , 0 ) self.assertListEqual(dset.column_names , [] )
83
1
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _SCREAMING_SNAKE_CASE = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def snake_case ( snake_case__ :Union[str, Any]) -> Dict: _A = torch.load(snake_case__ , map_location="""cpu""") return sd def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any] , snake_case__ :int=rename_keys_prefix) -> Optional[Any]: _A = OrderedDict() _A = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _A = key for name_pair in rename_keys_prefix: _A = new_key.replace(name_pair[0] , name_pair[1]) _A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _A = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple) -> int: assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _A = """pretraining""" if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "nlvr" in checkpoint_path: _A = {"""visual_embedding_dim""": 1_024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} _A = """multichoice""" elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} _A = """vqa_advanced""" elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048, """num_labels""": 3_129} _A = """vqa""" elif "nlvr" in checkpoint_path: _A = { """visual_embedding_dim""": 1_024, """num_labels""": 2, } _A = """nlvr""" _A = VisualBertConfig(**snake_case__) # Load State Dict _A = load_state_dict(snake_case__) _A = get_new_dict(snake_case__ , snake_case__) if model_type == "pretraining": _A = VisualBertForPreTraining(snake_case__) elif model_type == "vqa": _A = VisualBertForQuestionAnswering(snake_case__) elif model_type == "nlvr": _A = VisualBertForVisualReasoning(snake_case__) elif model_type == "multichoice": _A = VisualBertForMultipleChoice(snake_case__) model.load_state_dict(snake_case__) # Save Checkpoints Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
83
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case__ , snake_case__))) _A = [float(snake_case__) for n in range(limit + 1)] for p in primes: for n in range(snake_case__ , limit + 1 , snake_case__): phi[n] *= 1 - 1 / p return int(sum(phi[2:])) if __name__ == "__main__": print(F'''{solution() = }''')
83
1
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def snake_case ( snake_case__ :List[Any] , snake_case__ :Optional[Any]) -> Dict: _A = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg""" _A = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("""RGB""") _A = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC), transforms.ToTensor(), transforms.Normalize((0.4814_5466, 0.457_8275, 0.4082_1073) , (0.2686_2954, 0.2613_0258, 0.2757_7711)), ]) _A = transform(snake_case__).unsqueeze(0).to(snake_case__) return image def snake_case ( snake_case__ :Optional[int]) -> List[Any]: if "visual_encoder" in key: _A = re.sub("""visual_encoder*""" , """vision_model.encoder""" , snake_case__) if "blocks" in key: _A = re.sub(R"""blocks""" , """layers""" , snake_case__) if "attn" in key: _A = re.sub(R"""attn""" , """self_attn""" , snake_case__) if "norm1" in key: _A = re.sub(R"""norm1""" , """layer_norm1""" , snake_case__) if "norm2" in key: _A = re.sub(R"""norm2""" , """layer_norm2""" , snake_case__) if "encoder.norm" in key: _A = re.sub(R"""encoder.norm""" , """post_layernorm""" , snake_case__) if "encoder.patch_embed.proj" in key: _A = re.sub(R"""encoder.patch_embed.proj""" , """embeddings.patch_embedding""" , snake_case__) if "encoder.pos_embed" in key: _A = re.sub(R"""encoder.pos_embed""" , """embeddings.position_embedding""" , snake_case__) if "encoder.cls_token" in key: _A = re.sub(R"""encoder.cls_token""" , """embeddings.class_embedding""" , snake_case__) if "self_attn" in key: _A = re.sub(R"""self_attn.proj""" , """self_attn.projection""" , snake_case__) return key @torch.no_grad() def snake_case ( snake_case__ :int , snake_case__ :Any=None) -> Any: if config_path is not None: _A = BlipConfig.from_pretrained(snake_case__) else: _A = BlipConfig(projection_dim=512 , text_config={} , vision_config={}) _A = BlipForConditionalGeneration(snake_case__).eval() _A = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth""" _A = blip_decoder(pretrained=snake_case__ , image_size=384 , vit="""base""") _A = pt_model.eval() _A = pt_model.state_dict() for key in modified_state_dict.copy(): _A = modified_state_dict.pop(snake_case__) _A = rename_key(snake_case__) _A = value hf_model.load_state_dict(snake_case__) _A = 384 _A = load_demo_image(image_size=snake_case__ , device="""cpu""") _A = BertTokenizer.from_pretrained("""bert-base-uncased""") _A = tokenizer(["""a picture of"""]).input_ids _A = hf_model.generate(snake_case__ , snake_case__) assert out[0].tolist() == [30_522, 1_037, 3_861, 1_997, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] _A = hf_model.generate(snake_case__) assert out[0].tolist() == [30_522, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(snake_case__) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' _A = ( """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth""" ) _A = blip_vqa(pretrained=snake_case__ , image_size=snake_case__ , vit="""base""") vqa_model.eval() _A = vqa_model.state_dict() for key in modified_state_dict.copy(): _A = modified_state_dict.pop(snake_case__) _A = rename_key(snake_case__) _A = value _A = BlipForQuestionAnswering(snake_case__) hf_vqa_model.load_state_dict(snake_case__) _A = ["""How many dogs are in this image?"""] _A = tokenizer(snake_case__ , return_tensors="""pt""").input_ids _A = hf_vqa_model.generate(snake_case__ , snake_case__) print(tokenizer.decode(answer[0])) assert tokenizer.decode(answer[0]) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + """_vqa""") _A = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth""" _A = blip_itm(pretrained=snake_case__ , image_size=snake_case__ , vit="""base""") itm_model.eval() _A = itm_model.state_dict() for key in modified_state_dict.copy(): _A = modified_state_dict.pop(snake_case__) _A = rename_key(snake_case__) _A = value _A = BlipForImageTextRetrieval(snake_case__) _A = ["""A picture of a woman with a dog sitting in a beach"""] _A = tokenizer( snake_case__ , return_tensors="""pt""" , padding="""max_length""" , truncation=snake_case__ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(snake_case__) hf_itm_model.eval() _A = hf_itm_model(snake_case__ , snake_case__ , use_itm_head=snake_case__) _A = hf_itm_model(snake_case__ , snake_case__ , use_itm_head=snake_case__) assert out[0].item() == 0.2110_6874_9427_7954 assert torch.nn.functional.softmax(out_itm[0] , dim=1)[:, 1].item() == 0.4_5698_8453_8650_5127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + """_itm""") if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
83
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=32 , lowerCAmelCase_=5 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_="None" , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=None , ) -> Union[str, Any]: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = num_choices _A = relative_attention _A = position_biased_input _A = pos_att_type _A = scope def UpperCAmelCase ( self ) -> Dict: _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self ) -> Optional[int]: return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Any: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: _A = DebertaVaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = DebertaVaForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: _A = self.num_labels _A = DebertaVaForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = self.num_labels _A = DebertaVaForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _A = DebertaVaForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = DebertaVaForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) lowerCamelCase :str = ( { '''feature-extraction''': DebertaVaModel, '''fill-mask''': DebertaVaForMaskedLM, '''question-answering''': DebertaVaForQuestionAnswering, '''text-classification''': DebertaVaForSequenceClassification, '''token-classification''': DebertaVaForTokenClassification, '''zero-shot''': DebertaVaForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase :str = True lowerCamelCase :Union[str, Any] = False lowerCamelCase :Optional[int] = False lowerCamelCase :List[str] = False lowerCamelCase :str = False def UpperCAmelCase ( self ) -> Optional[int]: _A = DebertaVaModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> List[str]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> List[str]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> Any: for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = DebertaVaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason="""Model not available yet""" ) def UpperCAmelCase ( self ) -> int: pass @slow def UpperCAmelCase ( self ) -> Optional[Any]: _A = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _A = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) _A = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] # compare the actual values for a slice. _A = torch.tensor( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
83
1
import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Optional[int] = ['''image_processor''', '''tokenizer'''] lowerCamelCase :str = '''BlipImageProcessor''' lowerCamelCase :Optional[Any] = '''AutoTokenizer''' def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: super().__init__(lowerCAmelCase_ , lowerCAmelCase_ ) # add QFormer tokenizer _A = qformer_tokenizer def __call__( self , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = True , lowerCAmelCase_ = False , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = 0 , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = False , lowerCAmelCase_ = False , lowerCAmelCase_ = False , lowerCAmelCase_ = False , lowerCAmelCase_ = False , lowerCAmelCase_ = True , lowerCAmelCase_ = None , **lowerCAmelCase_ , ) -> BatchFeature: if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) _A = BatchFeature() if text is not None: _A = self.tokenizer( text=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , stride=lowerCAmelCase_ , pad_to_multiple_of=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , return_overflowing_tokens=lowerCAmelCase_ , return_special_tokens_mask=lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_length=lowerCAmelCase_ , verbose=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ , ) encoding.update(lowerCAmelCase_ ) _A = self.qformer_tokenizer( text=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , stride=lowerCAmelCase_ , pad_to_multiple_of=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , return_overflowing_tokens=lowerCAmelCase_ , return_special_tokens_mask=lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_length=lowerCAmelCase_ , verbose=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A = qformer_text_encoding.pop("""input_ids""" ) _A = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: _A = self.image_processor(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ ) encoding.update(lowerCAmelCase_ ) return encoding def UpperCAmelCase ( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Dict: return self.tokenizer.batch_decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Any: return self.tokenizer.decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase ( self ) -> Any: _A = self.tokenizer.model_input_names _A = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCAmelCase ( self , lowerCAmelCase_ , **lowerCAmelCase_ ) -> Dict: if os.path.isfile(lowerCAmelCase_ ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) _A = os.path.join(lowerCAmelCase_ , """qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(lowerCAmelCase_ ) return super().save_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) @classmethod def UpperCAmelCase ( cls , lowerCAmelCase_ , **lowerCAmelCase_ ) -> Dict: _A = AutoTokenizer.from_pretrained(lowerCAmelCase_ , subfolder="""qformer_tokenizer""" ) _A = cls._get_arguments_from_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) args.append(lowerCAmelCase_ ) return cls(*lowerCAmelCase_ )
83
def snake_case ( snake_case__ :int , snake_case__ :int) -> int: return int(input_a == input_a == 0) def snake_case ( ) -> None: print("""Truth Table of NOR Gate:""") print("""| Input 1 | Input 2 | Output |""") print(F'''| 0 | 0 | {nor_gate(0 , 0)} |''') print(F'''| 0 | 1 | {nor_gate(0 , 1)} |''') print(F'''| 1 | 0 | {nor_gate(1 , 0)} |''') print(F'''| 1 | 1 | {nor_gate(1 , 1)} |''') if __name__ == "__main__": import doctest doctest.testmod() main()
83
1
import random def snake_case ( snake_case__ :Union[str, Any] , snake_case__ :Tuple , snake_case__ :List[Any]) -> int: _A = a[left_index] _A = left_index + 1 for j in range(left_index + 1 , snake_case__): if a[j] < pivot: _A , _A = a[i], a[j] i += 1 _A , _A = a[i - 1], a[left_index] return i - 1 def snake_case ( snake_case__ :Tuple , snake_case__ :Optional[Any] , snake_case__ :List[str]) -> Union[str, Any]: if left < right: _A = random.randint(snake_case__ , right - 1) _A , _A = ( a[left], a[pivot], ) # switches the pivot with the left most bound _A = partition(snake_case__ , snake_case__ , snake_case__) quick_sort_random( snake_case__ , snake_case__ , snake_case__) # recursive quicksort to the left of the pivot point quick_sort_random( snake_case__ , pivot_index + 1 , snake_case__) # recursive quicksort to the right of the pivot point def snake_case ( ) -> Optional[int]: _A = input("""Enter numbers separated by a comma:\n""").strip() _A = [int(snake_case__) for item in user_input.split(""",""")] quick_sort_random(snake_case__ , 0 , len(snake_case__)) print(snake_case__) if __name__ == "__main__": main()
83
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=sys.maxsize ) -> str: _A = """bilinear""" _A = max_size _A = short_edge_length def __call__( self , lowerCAmelCase_ ) -> Optional[Any]: _A = [] for img in imgs: _A , _A = img.shape[:2] # later: provide list and randomly choose index for resize _A = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img _A = size * 1.0 / min(lowerCAmelCase_ , lowerCAmelCase_ ) if h < w: _A , _A = size, scale * w else: _A , _A = scale * h, size if max(lowerCAmelCase_ , lowerCAmelCase_ ) > self.max_size: _A = self.max_size * 1.0 / max(lowerCAmelCase_ , lowerCAmelCase_ ) _A = newh * scale _A = neww * scale _A = int(neww + 0.5 ) _A = int(newh + 0.5 ) if img.dtype == np.uinta: _A = Image.fromarray(lowerCAmelCase_ ) _A = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) _A = np.asarray(lowerCAmelCase_ ) else: _A = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw _A = nn.functional.interpolate( lowerCAmelCase_ , (newh, neww) , mode=self.interp_method , align_corners=lowerCAmelCase_ ).squeeze(0 ) img_augs.append(lowerCAmelCase_ ) return img_augs class a : """simple docstring""" def __init__( self , lowerCAmelCase_ ) -> List[Any]: _A = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) _A = cfg.INPUT.FORMAT _A = cfg.SIZE_DIVISIBILITY _A = cfg.PAD_VALUE _A = cfg.INPUT.MAX_SIZE_TEST _A = cfg.MODEL.DEVICE _A = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = lambda lowerCAmelCase_ : (x - self.pixel_mean) / self.pixel_std def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = tuple(max(lowerCAmelCase_ ) for s in zip(*[img.shape for img in images] ) ) _A = [im.shape[-2:] for im in images] _A = [ nn.functional.pad( lowerCAmelCase_ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] return torch.stack(lowerCAmelCase_ ), torch.tensor(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_=False ) -> int: with torch.no_grad(): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = [images] if single_image: assert len(lowerCAmelCase_ ) == 1 for i in range(len(lowerCAmelCase_ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(lowerCAmelCase_ , images.pop(lowerCAmelCase_ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( lowerCAmelCase_ , torch.as_tensor(img_tensorize(images.pop(lowerCAmelCase_ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge _A = torch.tensor([im.shape[:2] for im in images] ) _A = self.aug(lowerCAmelCase_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic _A = [self.normalizer(lowerCAmelCase_ ) for x in images] # now pad them to do the following operations _A , _A = self.pad(lowerCAmelCase_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad _A = torch.true_divide(lowerCAmelCase_ , lowerCAmelCase_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[Any]) -> Tuple: boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def snake_case ( snake_case__ :Optional[int] , snake_case__ :Tuple[int, int]) -> Optional[Any]: assert torch.isfinite(snake_case__).all(), "Box tensor contains infinite or NaN!" _A , _A = box_size tensor[:, 0].clamp_(min=0 , max=snake_case__) tensor[:, 1].clamp_(min=0 , max=snake_case__) tensor[:, 2].clamp_(min=0 , max=snake_case__) tensor[:, 3].clamp_(min=0 , max=snake_case__)
83
1
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def UpperCAmelCase ( self ) -> Optional[int]: _A = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCAmelCase_ ): self.assertDictEqual(lowerCAmelCase_ , example_records[i] ) def UpperCAmelCase ( self ) -> str: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) _A = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def UpperCAmelCase ( self ) -> Any: # checks what happens with missing columns _A = [{"""col_1""": 1}, {"""col_2""": """x"""}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def UpperCAmelCase ( self ) -> Tuple: # checks if the type can be inferred from the second record _A = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def UpperCAmelCase ( self ) -> Any: _A = Dataset.from_list([] ) self.assertEqual(len(lowerCAmelCase_ ) , 0 ) self.assertListEqual(dset.column_names , [] )
83
from collections import defaultdict def snake_case ( snake_case__ :int) -> int: _A = 1 _A = True for v in tree[start]: if v not in visited: ret += dfs(snake_case__) if ret % 2 == 0: cuts.append(snake_case__) return ret def snake_case ( ) -> Any: dfs(1) if __name__ == "__main__": _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10, 9 _SCREAMING_SNAKE_CASE = defaultdict(list) _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
83
1
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def snake_case ( snake_case__ :str) -> str: if is_torch_version("""<""" , """2.0.0""") or not hasattr(snake_case__ , """_dynamo"""): return False return isinstance(snake_case__ , torch._dynamo.eval_frame.OptimizedModule) def snake_case ( snake_case__ :Union[str, Any] , snake_case__ :bool = True) -> List[str]: _A = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _A = is_compiled_module(snake_case__) if is_compiled: _A = model _A = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(snake_case__ , snake_case__): _A = model.module if not keep_fpaa_wrapper: _A = getattr(snake_case__ , """forward""") _A = model.__dict__.pop("""_original_forward""" , snake_case__) if original_forward is not None: while hasattr(snake_case__ , """__wrapped__"""): _A = forward.__wrapped__ if forward == original_forward: break _A = forward if getattr(snake_case__ , """_converted_to_transformer_engine""" , snake_case__): convert_model(snake_case__ , to_transformer_engine=snake_case__) if is_compiled: _A = model _A = compiled_model return model def snake_case ( ) -> List[Any]: PartialState().wait_for_everyone() def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[int]) -> Any: if PartialState().distributed_type == DistributedType.TPU: xm.save(snake_case__ , snake_case__) elif PartialState().local_process_index == 0: torch.save(snake_case__ , snake_case__) @contextmanager def snake_case ( **snake_case__ :Tuple) -> str: for key, value in kwargs.items(): _A = str(snake_case__) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def snake_case ( snake_case__ :Any) -> List[str]: if not hasattr(snake_case__ , """__qualname__""") and not hasattr(snake_case__ , """__name__"""): _A = getattr(snake_case__ , """__class__""" , snake_case__) if hasattr(snake_case__ , """__qualname__"""): return obj.__qualname__ if hasattr(snake_case__ , """__name__"""): return obj.__name__ return str(snake_case__) def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[int]) -> List[Any]: for key, value in source.items(): if isinstance(snake_case__ , snake_case__): _A = destination.setdefault(snake_case__ , {}) merge_dicts(snake_case__ , snake_case__) else: _A = value return destination def snake_case ( snake_case__ :int = None) -> bool: if port is None: _A = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM) as s: return s.connect_ex(("""localhost""", port)) == 0
83
import heapq def snake_case ( snake_case__ :dict) -> set[int]: _A = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(snake_case__ , [-1 * len(snake_case__), (key, value)]) # chosen_vertices = set of chosen vertices _A = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _A = heapq.heappop(snake_case__)[1][0] chosen_vertices.add(snake_case__) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _A = elem[1][1].index(snake_case__) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(snake_case__) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(F'''Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}''')
83
1
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :torch.FloatTensor lowerCamelCase :torch.FloatTensor class a ( __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Dict = 1 @register_to_config def __init__( self , lowerCAmelCase_ = 20_00 , lowerCAmelCase_ = 0.15 , lowerCAmelCase_ = 0.01 , lowerCAmelCase_ = 1348.0 , lowerCAmelCase_ = 1E-5 , lowerCAmelCase_ = 1 , ) -> str: # standard deviation of the initial noise distribution _A = sigma_max # setable values _A = None self.set_sigmas(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> torch.FloatTensor: return sample def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None ) -> Optional[Any]: _A = sampling_eps if sampling_eps is not None else self.config.sampling_eps _A = torch.linspace(1 , lowerCAmelCase_ , lowerCAmelCase_ , device=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None ) -> int: _A = sigma_min if sigma_min is not None else self.config.sigma_min _A = sigma_max if sigma_max is not None else self.config.sigma_max _A = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(lowerCAmelCase_ , lowerCAmelCase_ ) _A = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _A = torch.exp(torch.linspace(math.log(lowerCAmelCase_ ) , math.log(lowerCAmelCase_ ) , lowerCAmelCase_ ) ) _A = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = True , ) -> Union[SdeVeOutput, Tuple]: if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler""" ) _A = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _A = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _A = timesteps.to(self.discrete_sigmas.device ) _A = self.discrete_sigmas[timesteps].to(sample.device ) _A = self.get_adjacent_sigma(lowerCAmelCase_ , lowerCAmelCase_ ).to(sample.device ) _A = torch.zeros_like(lowerCAmelCase_ ) _A = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _A = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _A = diffusion.unsqueeze(-1 ) _A = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _A = randn_tensor( sample.shape , layout=sample.layout , generator=lowerCAmelCase_ , device=sample.device , dtype=sample.dtype ) _A = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _A = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=lowerCAmelCase_ , prev_sample_mean=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = True , ) -> Union[SchedulerOutput, Tuple]: if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler""" ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _A = randn_tensor(sample.shape , layout=sample.layout , generator=lowerCAmelCase_ ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _A = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _A = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _A = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _A = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _A = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _A = step_size.unsqueeze(-1 ) _A = sample + step_size * model_output _A = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) -> torch.FloatTensor: # Make sure sigmas and timesteps have the same device and dtype as original_samples _A = timesteps.to(original_samples.device ) _A = self.discrete_sigmas.to(original_samples.device )[timesteps] _A = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(lowerCAmelCase_ ) * sigmas[:, None, None, None] ) _A = noise + original_samples return noisy_samples def __len__( self ) -> Tuple: return self.config.num_train_timesteps
83
import math import unittest def snake_case ( snake_case__ :int) -> bool: assert isinstance(snake_case__ , snake_case__) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def UpperCAmelCase ( self ) -> Dict: with self.assertRaises(lowerCAmelCase_ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , """Zero doesn't have any positive factors, primes must have exactly two.""" , ) self.assertFalse( is_prime(1 ) , """One only has 1 positive factor, primes must have exactly two.""" , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
83
1
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Union[str, Any]: _A = inspect.getfile(accelerate.test_utils ) _A = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """external_deps""", """test_metrics.py"""] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 _A = test_metrics @require_cpu def UpperCAmelCase ( self ) -> Any: debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def UpperCAmelCase ( self ) -> str: debug_launcher(self.test_metrics.main ) @require_single_gpu def UpperCAmelCase ( self ) -> Dict: self.test_metrics.main() @require_multi_gpu def UpperCAmelCase ( self ) -> str: print(F'''Found {torch.cuda.device_count()} devices.''' ) _A = ["""torchrun""", F'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowerCAmelCase_ , env=os.environ.copy() )
83
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device _SCREAMING_SNAKE_CASE = False class a ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Tuple: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self ) -> Dict: _A = VersatileDiffusionTextToImagePipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = """A painting of a squirrel eating a burger """ _A = torch.manual_seed(0 ) _A = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _A = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = generator.manual_seed(0 ) _A = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def UpperCAmelCase ( self ) -> List[Any]: _A = VersatileDiffusionTextToImagePipeline.from_pretrained( """shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = """A painting of a squirrel eating a burger """ _A = torch.manual_seed(0 ) _A = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images _A = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _A = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
83
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(__lowerCAmelCase ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Optional[Any]: super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) self.check_model_type(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , **lowerCAmelCase_ ) -> Tuple: _A , _A = {}, {} if padding is not None: _A = padding if truncation is not None: _A = truncation if top_k is not None: _A = top_k return preprocess_params, {}, postprocess_params def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , **lowerCAmelCase_ ) -> Union[str, Any]: if isinstance(lowerCAmelCase_ , (Image.Image, str) ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = {"""image""": image, """question""": question} else: _A = image _A = super().__call__(lowerCAmelCase_ , **lowerCAmelCase_ ) return results def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Any: _A = load_image(inputs["""image"""] ) _A = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ ) _A = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase_ ) return model_inputs def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = self.model(**lowerCAmelCase_ ) return model_outputs def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=5 ) -> Union[str, Any]: if top_k > self.model.config.num_labels: _A = self.model.config.num_labels if self.framework == "pt": _A = model_outputs.logits.sigmoid()[0] _A , _A = probs.topk(lowerCAmelCase_ ) else: raise ValueError(F'''Unsupported framework: {self.framework}''' ) _A = scores.tolist() _A = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowerCAmelCase_ , lowerCAmelCase_ )]
83
1
from __future__ import annotations def snake_case ( snake_case__ :float , snake_case__ :float , snake_case__ :float) -> dict[str, float]: if (voltage, current, resistance).count(0) != 1: raise ValueError("""One and only one argument must be 0""") if resistance < 0: raise ValueError("""Resistance cannot be negative""") if voltage == 0: return {"voltage": float(current * resistance)} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("""Exactly one argument must be 0""") if __name__ == "__main__": import doctest doctest.testmod()
83
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :str , snake_case__ :PreTrainedTokenizer , snake_case__ :int , snake_case__ :Optional[int] = None , ) -> Optional[int]: _A = {} if train_file is not None: _A = [train_file] if eval_file is not None: _A = [eval_file] if test_file is not None: _A = [test_file] _A = datasets.load_dataset("""csv""" , data_files=snake_case__) _A = list(ds[list(files.keys())[0]].features.keys()) _A = features_name.pop(snake_case__) _A = list(set(ds[list(files.keys())[0]][label_name])) _A = {label: i for i, label in enumerate(snake_case__)} _A = tokenizer.model_input_names _A = {} if len(snake_case__) == 1: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( example[features_name[0]] , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""") , batched=snake_case__ , ) elif len(snake_case__) == 2: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""" , ) , batched=snake_case__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _A = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _A = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _A = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST]))) return train_ds, val_ds, test_ds, labelaid _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class a : """simple docstring""" lowerCamelCase :int = field(metadata={'''help''': '''Which column contains the label'''} ) lowerCamelCase :str = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the training file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the development file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the test file'''} ) lowerCamelCase :int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCamelCase :bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) @dataclass class a : """simple docstring""" lowerCamelCase :str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) lowerCamelCase :bool = field(default=__lowerCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) def snake_case ( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments)) _A , _A , _A = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir) and os.listdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""") # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info( F'''n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1)}, ''' F'''16-bits training: {training_args.fpaa}''') logger.info(F'''Training/evaluation parameters {training_args}''') # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _A = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _A , _A , _A , _A = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=snake_case__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _A = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(snake_case__) , labelaid=snake_case__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="""text-classification""" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _A = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(""".bin""" in model_args.model_name_or_path) , config=snake_case__ , cache_dir=model_args.cache_dir , ) def compute_metrics(snake_case__ :EvalPrediction) -> Dict: _A = np.argmax(p.predictions , axis=1) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _A = TFTrainer( model=snake_case__ , args=snake_case__ , train_dataset=snake_case__ , eval_dataset=snake_case__ , compute_metrics=snake_case__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir) # Evaluation _A = {} if training_args.do_eval: logger.info("""*** Evaluate ***""") _A = trainer.evaluate() _A = os.path.join(training_args.output_dir , """eval_results.txt""") with open(snake_case__ , """w""") as writer: logger.info("""***** Eval results *****""") for key, value in result.items(): logger.info(F''' {key} = {value}''') writer.write(F'''{key} = {value}\n''') results.update(snake_case__) return results if __name__ == "__main__": main()
83
1
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
83
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = '''speech_to_text''' lowerCamelCase :List[str] = ['''past_key_values'''] lowerCamelCase :str = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=12 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=2 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=60_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=2 , lowerCAmelCase_=(5, 5) , lowerCAmelCase_=10_24 , lowerCAmelCase_=80 , lowerCAmelCase_=1 , **lowerCAmelCase_ , ) -> Tuple: _A = vocab_size _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = encoder_layerdrop _A = decoder_layerdrop _A = use_cache _A = encoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = max_source_positions _A = max_target_positions _A = num_conv_layers _A = list(lowerCAmelCase_ ) _A = conv_channels _A = input_feat_per_channel _A = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` """ F'''but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, ''' F'''`config.num_conv_layers = {self.num_conv_layers}`.''' ) super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , )
83
1
_SCREAMING_SNAKE_CASE = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def snake_case ( snake_case__ :Any , snake_case__ :Union[str, Any] , snake_case__ :Any , snake_case__ :List[Any]) -> Optional[Any]: # Return True if there is node that has not iterated. _A = [False] * len(snake_case__) _A = [s] _A = True while queue: _A = queue.pop(0) for ind in range(len(graph[u])): if visited[ind] is False and graph[u][ind] > 0: queue.append(snake_case__) _A = True _A = u return visited[t] def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :Union[str, Any]) -> int: _A = [-1] * (len(snake_case__)) _A = 0 _A = [] _A = [i[:] for i in graph] # Record original cut, copy. while bfs(snake_case__ , snake_case__ , snake_case__ , snake_case__): _A = float("""Inf""") _A = sink while s != source: # Find the minimum value in select path _A = min(snake_case__ , graph[parent[s]][s]) _A = parent[s] max_flow += path_flow _A = sink while v != source: _A = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow _A = parent[v] for i in range(len(snake_case__)): for j in range(len(graph[0])): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j)) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
83
from __future__ import annotations from collections.abc import Callable def snake_case ( snake_case__ :Callable[[int | float], int | float] , snake_case__ :int | float , snake_case__ :int | float , snake_case__ :int = 100 , ) -> float: _A = x_start _A = fnc(snake_case__) _A = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area _A = (x_end - x_start) / steps + xa _A = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step _A = xa _A = fxa return area if __name__ == "__main__": def snake_case ( snake_case__ :Tuple) -> List[str]: return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') _SCREAMING_SNAKE_CASE = 10 while i <= 100_000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
83
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = '▁' _SCREAMING_SNAKE_CASE = {'vocab_file': 'sentencepiece.bpe.model'} _SCREAMING_SNAKE_CASE = { 'vocab_file': { 'facebook/mbart-large-50-one-to-many-mmt': ( 'https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model' ), } } _SCREAMING_SNAKE_CASE = { 'facebook/mbart-large-50-one-to-many-mmt': 1_024, } # fmt: off _SCREAMING_SNAKE_CASE = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN', 'af_ZA', 'az_AZ', 'bn_IN', 'fa_IR', 'he_IL', 'hr_HR', 'id_ID', 'ka_GE', 'km_KH', 'mk_MK', 'ml_IN', 'mn_MN', 'mr_IN', 'pl_PL', 'ps_AF', 'pt_XX', 'sv_SE', 'sw_KE', 'ta_IN', 'te_IN', 'th_TH', 'tl_XX', 'uk_UA', 'ur_PK', 'xh_ZA', 'gl_ES', 'sl_SI'] class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :List[str] = VOCAB_FILES_NAMES lowerCamelCase :Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase :int = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase :List[Any] = ['''input_ids''', '''attention_mask'''] lowerCamelCase :List[int] = [] lowerCamelCase :List[int] = [] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_="</s>" , lowerCAmelCase_="</s>" , lowerCAmelCase_="<s>" , lowerCAmelCase_="<unk>" , lowerCAmelCase_="<pad>" , lowerCAmelCase_="<mask>" , lowerCAmelCase_ = None , **lowerCAmelCase_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it _A = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else mask_token _A = {} if sp_model_kwargs is None else sp_model_kwargs _A = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=lowerCAmelCase_ , tgt_lang=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase_ , ) _A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCAmelCase_ ) ) _A = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _A = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _A = 1 _A = len(self.sp_model ) _A = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase_ ) } _A = {v: k for k, v in self.lang_code_to_id.items()} _A = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _A = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _A = src_lang if src_lang is not None else """en_XX""" _A = self.lang_code_to_id[self._src_lang] _A = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCAmelCase ( self ) -> int: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def UpperCAmelCase ( self ) -> str: return self._src_lang @src_lang.setter def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ) -> Dict: _A = self.__dict__.copy() _A = None return state def __setstate__( self , lowerCAmelCase_ ) -> None: _A = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _A = {} _A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase ( self ) -> Dict: _A = {self.convert_ids_to_tokens(lowerCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[str]: return self.sp_model.encode(lowerCAmelCase_ , out_type=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _A = self.sp_model.PieceToId(lowerCAmelCase_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = [] _A = """""" _A = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCAmelCase_ ) + token _A = True _A = [] else: current_sub_tokens.append(lowerCAmelCase_ ) _A = False out_string += self.sp_model.decode(lowerCAmelCase_ ) return out_string.strip() def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase_ , """wb""" ) as fi: _A = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase_ ) return (out_vocab_file,) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase_ , token_ids_a=lowerCAmelCase_ , already_has_special_tokens=lowerCAmelCase_ ) _A = [1] * len(self.prefix_tokens ) _A = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase_ )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase_ )) + ([0] * len(lowerCAmelCase_ )) + suffix_ones def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> Tuple: if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _A = src_lang _A = self(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ ) _A = self.convert_tokens_to_ids(lowerCAmelCase_ ) _A = tgt_lang_id return inputs def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = "en_XX" , lowerCAmelCase_ = None , lowerCAmelCase_ = "ro_RO" , **lowerCAmelCase_ , ) -> BatchEncoding: _A = src_lang _A = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: return self.set_src_lang_special_tokens(self.src_lang ) def UpperCAmelCase ( self ) -> List[str]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = self.lang_code_to_id[src_lang] _A = [self.cur_lang_code_id] _A = [self.eos_token_id] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = self.lang_code_to_id[tgt_lang] _A = [self.cur_lang_code_id] _A = [self.eos_token_id]
83
import numpy as np import qiskit def snake_case ( snake_case__ :int = 8 , snake_case__ :int | None = None) -> str: _A = np.random.default_rng(seed=snake_case__) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _A = 6 * key_len # Measurement basis for Alice's qubits. _A = rng.integers(2 , size=snake_case__) # The set of states Alice will prepare. _A = rng.integers(2 , size=snake_case__) # Measurement basis for Bob's qubits. _A = rng.integers(2 , size=snake_case__) # Quantum Circuit to simulate BB84 _A = qiskit.QuantumCircuit(snake_case__ , name="""BB84""") # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case__): if alice_state[index] == 1: bbaa_circ.x(snake_case__) if alice_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case__): if bob_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _A = qiskit.Aer.get_backend("""aer_simulator""") # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _A = qiskit.execute(snake_case__ , snake_case__ , shots=1 , seed_simulator=snake_case__) # Returns the result of measurement. _A = job.result().get_counts(snake_case__).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _A = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case__ , snake_case__ , snake_case__) if alice_basis_bit == bob_basis_bit ]) # Get final key. Pad with 0 if too short, otherwise truncate. _A = gen_key[:key_len] if len(snake_case__) >= key_len else gen_key.ljust(snake_case__ , """0""") return key if __name__ == "__main__": print(F'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
83
1
from __future__ import annotations import math def snake_case ( snake_case__ :int) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True _SCREAMING_SNAKE_CASE = [num for num in range(3, 100_001, 2) if not is_prime(num)] def snake_case ( snake_case__ :int) -> list[int]: if not isinstance(snake_case__ , snake_case__): raise ValueError("""n must be an integer""") if n <= 0: raise ValueError("""n must be >= 0""") _A = [] for num in range(len(snake_case__)): _A = 0 while 2 * i * i <= odd_composites[num]: _A = odd_composites[num] - 2 * i * i if is_prime(snake_case__): break i += 1 else: list_nums.append(odd_composites[num]) if len(snake_case__) == n: return list_nums return [] def snake_case ( ) -> int: return compute_nums(1)[0] if __name__ == "__main__": print(F'''{solution() = }''')
83
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def snake_case ( snake_case__ :int) -> Optional[int]: return EnvironmentCommand() def snake_case ( snake_case__ :Tuple) -> List[str]: return EnvironmentCommand(args.accelerate_config_file) class a ( __lowerCAmelCase ): """simple docstring""" @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: _A = parser.add_parser("""env""" ) download_parser.set_defaults(func=lowerCAmelCase_ ) download_parser.add_argument( """--accelerate-config_file""" , default=lowerCAmelCase_ , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=lowerCAmelCase_ ) def __init__( self , lowerCAmelCase_ , *lowerCAmelCase_ ) -> None: _A = accelerate_config_file def UpperCAmelCase ( self ) -> Dict: _A = """not installed""" if is_safetensors_available(): import safetensors _A = safetensors.__version__ elif importlib.util.find_spec("""safetensors""" ) is not None: import safetensors _A = F'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' _A = """not installed""" _A = _A = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file _A = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(lowerCAmelCase_ ): _A = load_config_from_file(self._accelerate_config_file ).to_dict() _A = ( """\n""".join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else F'''\t{accelerate_config}''' ) _A = """not installed""" _A = """NA""" if is_torch_available(): import torch _A = torch.__version__ _A = torch.cuda.is_available() _A = """not installed""" _A = """NA""" if is_tf_available(): import tensorflow as tf _A = tf.__version__ try: # deprecated in v2.1 _A = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool _A = bool(tf.config.list_physical_devices("""GPU""" ) ) _A = """not installed""" _A = """not installed""" _A = """not installed""" _A = """NA""" if is_flax_available(): import flax import jax import jaxlib _A = flax.__version__ _A = jax.__version__ _A = jaxlib.__version__ _A = jax.lib.xla_bridge.get_backend().platform _A = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": F'''{safetensors_version}''', """Accelerate version""": F'''{accelerate_version}''', """Accelerate config""": F'''{accelerate_config_str}''', """PyTorch version (GPU?)""": F'''{pt_version} ({pt_cuda_available})''', """Tensorflow version (GPU?)""": F'''{tf_version} ({tf_cuda_available})''', """Flax version (CPU?/GPU?/TPU?)""": F'''{flax_version} ({jax_backend})''', """Jax version""": F'''{jax_version}''', """JaxLib version""": F'''{jaxlib_version}''', """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""" ) print(self.format_dict(lowerCAmelCase_ ) ) return info @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: return "\n".join([F'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
83
1
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class a ( unittest.TestCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=32 , lowerCAmelCase_=5 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=4 , ) -> str: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_attention_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_choices def UpperCAmelCase ( self ) -> Any: _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_attention_mask: _A = random_attention_mask([self.batch_size, self.seq_length] ) _A = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCAmelCase_ , ) return config, input_ids, attention_mask def UpperCAmelCase ( self ) -> Tuple: _A = self.prepare_config_and_inputs() _A , _A , _A = config_and_inputs _A = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class a ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :Dict = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase ( self ) -> Any: _A = FlaxDistilBertModelTester(self ) @slow def UpperCAmelCase ( self ) -> Any: for model_class_name in self.all_model_classes: _A = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _A = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase_ ) @require_flax class a ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase ( self ) -> Union[str, Any]: _A = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _A = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _A = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] _A = (1, 11, 7_68) self.assertEqual(output.shape , lowerCAmelCase_ ) _A = np.array([[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1E-4 ) )
83
import colorsys from PIL import Image # type: ignore def snake_case ( snake_case__ :float , snake_case__ :float , snake_case__ :int) -> float: _A = x _A = y for step in range(snake_case__): # noqa: B007 _A = a * a - b * b + x _A = 2 * a * b + y _A = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255) for i in colorsys.hsv_to_rgb(snake_case__ , 1 , 1)) def snake_case ( snake_case__ :int = 800 , snake_case__ :int = 600 , snake_case__ :float = -0.6 , snake_case__ :float = 0 , snake_case__ :float = 3.2 , snake_case__ :int = 50 , snake_case__ :bool = True , ) -> Image.Image: _A = Image.new("""RGB""" , (image_width, image_height)) _A = img.load() # loop through the image-coordinates for image_x in range(snake_case__): for image_y in range(snake_case__): # determine the figure-coordinates based on the image-coordinates _A = figure_width / image_width * image_height _A = figure_center_x + (image_x / image_width - 0.5) * figure_width _A = figure_center_y + (image_y / image_height - 0.5) * figure_height _A = get_distance(snake_case__ , snake_case__ , snake_case__) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _A = get_color_coded_rgb(snake_case__) else: _A = get_black_and_white_rgb(snake_case__) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _SCREAMING_SNAKE_CASE = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
1
from __future__ import annotations from collections.abc import Callable def snake_case ( snake_case__ :Callable[[int | float], int | float] , snake_case__ :int | float , snake_case__ :int | float , snake_case__ :int = 100 , ) -> float: _A = x_start _A = fnc(snake_case__) _A = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area _A = (x_end - x_start) / steps + xa _A = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step _A = xa _A = fxa return area if __name__ == "__main__": def snake_case ( snake_case__ :Tuple) -> List[str]: return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') _SCREAMING_SNAKE_CASE = 10 while i <= 100_000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
83
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _SCREAMING_SNAKE_CASE = datasets.logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _SCREAMING_SNAKE_CASE = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _SCREAMING_SNAKE_CASE = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def snake_case ( snake_case__ :Optional[Any] , snake_case__ :str , snake_case__ :List[str]=False , snake_case__ :Dict=False , snake_case__ :Any=True , snake_case__ :List[str]=False , snake_case__ :Optional[Any]="dummy_doc") -> List[Any]: _A = {doc: key_lines} _A = {doc: sys_lines} _A = {} _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A , _A = reader.get_doc_mentions(snake_case__ , key_doc_lines[doc] , snake_case__) key_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) _A , _A = reader.get_doc_mentions(snake_case__ , sys_doc_lines[doc] , snake_case__) sys_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) if remove_nested: _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F'''annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}''') logger.info( """Number of resulting singleton clusters in the key """ F'''annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}''') if not keep_singletons: logger.info( F'''{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ''' """files, respectively""") return doc_coref_infos def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Dict , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Tuple) -> int: _A = get_coref_infos(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) _A = {} _A = 0 _A = 0 for name, metric in metrics: _A , _A , _A = evaluator.evaluate_documents(snake_case__ , snake_case__ , beta=1) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'''{name}/recall''': recall, F'''{name}/precision''': precision, F'''{name}/f1''': fa}) logger.info( name.ljust(10) , F'''Recall: {recall * 100:.2f}''' , F''' Precision: {precision * 100:.2f}''' , F''' F1: {fa * 100:.2f}''' , ) if conll_subparts_num == 3: _A = (conll / 3) * 100 logger.info(F'''CoNLL score: {conll:.2f}''') output_scores.update({"""conll_score""": conll}) return output_scores def snake_case ( snake_case__ :Union[str, Any]) -> List[Any]: _A = False for line in key_lines: if not line.startswith("""#"""): if len(line.split()) > 6: _A = line.split()[5] if not parse_col == "-": _A = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Union[str, Any]: _A = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _A = util.check_gold_parse_annotation(lowerCAmelCase_ ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _A = evaluate( key_lines=lowerCAmelCase_ , sys_lines=lowerCAmelCase_ , metrics=lowerCAmelCase_ , NP_only=lowerCAmelCase_ , remove_nested=lowerCAmelCase_ , keep_singletons=lowerCAmelCase_ , min_span=lowerCAmelCase_ , ) return score
83
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _SCREAMING_SNAKE_CASE = { 'configuration_maskformer': ['MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MaskFormerConfig'], 'configuration_maskformer_swin': ['MaskFormerSwinConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['MaskFormerFeatureExtractor'] _SCREAMING_SNAKE_CASE = ['MaskFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'MaskFormerForInstanceSegmentation', 'MaskFormerModel', 'MaskFormerPreTrainedModel', ] _SCREAMING_SNAKE_CASE = [ 'MaskFormerSwinBackbone', 'MaskFormerSwinModel', 'MaskFormerSwinPreTrainedModel', ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure)
83
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } _SCREAMING_SNAKE_CASE = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } _SCREAMING_SNAKE_CASE = {'facebook/blenderbot_small-90M': 512} def snake_case ( snake_case__ :Tuple) -> str: _A = set() _A = word[0] for char in word[1:]: pairs.add((prev_char, char)) _A = char _A = set(snake_case__) return pairs class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :List[Any] = VOCAB_FILES_NAMES lowerCamelCase :Tuple = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase :List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase :int = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="__start__" , lowerCAmelCase_="__end__" , lowerCAmelCase_="__unk__" , lowerCAmelCase_="__null__" , **lowerCAmelCase_ , ) -> int: super().__init__(unk_token=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , **lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) _A = {v: k for k, v in self.encoder.items()} with open(lowerCAmelCase_ , encoding="""utf-8""" ) as merges_handle: _A = merges_handle.read().split("""\n""" )[1:-1] _A = [tuple(merge.split() ) for merge in merges] _A = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_ ) ) ) ) _A = {} @property def UpperCAmelCase ( self ) -> int: return len(self.encoder ) def UpperCAmelCase ( self ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: if token in self.cache: return self.cache[token] _A = re.sub("""([.,!?()])""" , r""" \1""" , lowerCAmelCase_ ) _A = re.sub("""(')""" , r""" \1 """ , lowerCAmelCase_ ) _A = re.sub(r"""\s{2,}""" , """ """ , lowerCAmelCase_ ) if "\n" in token: _A = token.replace("""\n""" , """ __newln__""" ) _A = token.split(""" """ ) _A = [] for token in tokens: if not len(lowerCAmelCase_ ): continue _A = token.lower() _A = tuple(lowerCAmelCase_ ) _A = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) _A = get_pairs(lowerCAmelCase_ ) if not pairs: words.append(lowerCAmelCase_ ) continue while True: _A = min(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : self.bpe_ranks.get(lowerCAmelCase_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break _A , _A = bigram _A = [] _A = 0 while i < len(lowerCAmelCase_ ): try: _A = word.index(lowerCAmelCase_ , lowerCAmelCase_ ) new_word.extend(word[i:j] ) _A = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A = tuple(lowerCAmelCase_ ) _A = new_word if len(lowerCAmelCase_ ) == 1: break else: _A = get_pairs(lowerCAmelCase_ ) _A = """@@ """.join(lowerCAmelCase_ ) _A = word[:-4] _A = word words.append(lowerCAmelCase_ ) return " ".join(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[str]: _A = [] _A = re.findall(r"""\S+\n?""" , lowerCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(lowerCAmelCase_ ).split(""" """ ) ) ) return split_tokens def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: _A = token.lower() return self.encoder.get(lowerCAmelCase_ , self.encoder.get(self.unk_token ) ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: return self.decoder.get(lowerCAmelCase_ , self.unk_token ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = """ """.join(lowerCAmelCase_ ).replace("""@@ """ , """""" ).strip() return out_string def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase_ , ensure_ascii=lowerCAmelCase_ ) + """\n""" ) _A = 0 with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase_ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) _A = token_index writer.write(""" """.join(lowerCAmelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file
83
1
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html _SCREAMING_SNAKE_CASE = 'platform' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :int=None , snake_case__ :List[str]=None , snake_case__ :Dict=None , snake_case__ :Tuple=None , snake_case__ :Dict=None , snake_case__ :Optional[Any]=None , ) -> Tuple: if attention_mask is None: _A = np.where(input_ids != config.pad_token_id , 1 , 0) if decoder_attention_mask is None: _A = np.where(decoder_input_ids != config.pad_token_id , 1 , 0) if head_mask is None: _A = np.ones((config.encoder_layers, config.encoder_attention_heads)) if decoder_head_mask is None: _A = np.ones((config.decoder_layers, config.decoder_attention_heads)) if cross_attn_head_mask is None: _A = np.ones((config.decoder_layers, config.decoder_attention_heads)) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=99 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=4 , lowerCAmelCase_=4 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=32 , lowerCAmelCase_=2 , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=0.02 , ) -> Optional[int]: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = eos_token_id _A = pad_token_id _A = bos_token_id _A = initializer_range def UpperCAmelCase ( self ) -> Any: _A = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _A = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _A = shift_tokens_right(lowerCAmelCase_ , 1 , 2 ) _A = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCAmelCase_ , ) _A = prepare_blenderbot_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def UpperCAmelCase ( self ) -> Optional[int]: _A , _A = self.prepare_config_and_inputs() return config, inputs_dict def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _A = 20 _A = model_class_name(lowerCAmelCase_ ) _A = model.encode(inputs_dict["""input_ids"""] ) _A , _A = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _A = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_ ) _A = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _A = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _A = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) _A = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _A = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase_ , ) _A = model.decode(lowerCAmelCase_ , lowerCAmelCase_ ) _A = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _A = 20 _A = model_class_name(lowerCAmelCase_ ) _A = model.encode(inputs_dict["""input_ids"""] ) _A , _A = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _A = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _A = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_ ) _A = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _A = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) _A = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _A = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) _A = model.decode(lowerCAmelCase_ , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ ) _A = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) @require_flax class a ( unittest.TestCase ): """simple docstring""" lowerCamelCase :List[Any] = 99 def UpperCAmelCase ( self ) -> Tuple: _A = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _A = input_ids.shape[0] _A = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def UpperCAmelCase ( self ) -> List[Any]: _A , _A , _A = self._get_config_and_data() _A = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase_ ) _A = lm_model(input_ids=lowerCAmelCase_ ) _A = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _A = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase_ ) _A = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _A = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _A = lm_model(input_ids=lowerCAmelCase_ , decoder_input_ids=lowerCAmelCase_ ) _A = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _A = shift_tokens_right(lowerCAmelCase_ , 1 , 2 ) _A = np.equal(lowerCAmelCase_ , 1 ).astype(np.floataa ).sum() _A = np.equal(lowerCAmelCase_ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCAmelCase_ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class a ( __lowerCAmelCase , unittest.TestCase , __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = True lowerCamelCase :Optional[Any] = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCamelCase :List[str] = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def UpperCAmelCase ( self ) -> Any: _A = FlaxBlenderbotModelTester(self ) def UpperCAmelCase ( self ) -> Any: _A , _A = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: _A , _A = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _A = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) _A = model_class(lowerCAmelCase_ ) @jax.jit def encode_jitted(lowerCAmelCase_ , lowerCAmelCase_=None , **lowerCAmelCase_ ): return model.encode(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ) with self.subTest("""JIT Enabled""" ): _A = encode_jitted(**lowerCAmelCase_ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _A = encode_jitted(**lowerCAmelCase_ ).to_tuple() self.assertEqual(len(lowerCAmelCase_ ) , len(lowerCAmelCase_ ) ) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCAmelCase ( self ) -> Optional[int]: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _A = model_class(lowerCAmelCase_ ) _A = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) _A = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): return model.decode( decoder_input_ids=lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , encoder_outputs=lowerCAmelCase_ , ) with self.subTest("""JIT Enabled""" ): _A = decode_jitted(**lowerCAmelCase_ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _A = decode_jitted(**lowerCAmelCase_ ).to_tuple() self.assertEqual(len(lowerCAmelCase_ ) , len(lowerCAmelCase_ ) ) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCAmelCase ( self ) -> List[Any]: for model_class_name in self.all_model_classes: _A = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _A = np.ones((1, 1) ) * model.config.eos_token_id _A = model(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def UpperCAmelCase ( self ) -> Tuple: _A = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _A = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _A = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=lowerCAmelCase_ ) _A = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _A = ["""Sam"""] _A = tokenizer(lowerCAmelCase_ , return_tensors="""jax""" ) _A = model.generate(**lowerCAmelCase_ , **lowerCAmelCase_ ) _A = """Sam is a great name. It means \"sun\" in Gaelic.""" _A = tokenizer.batch_decode(lowerCAmelCase_ , **lowerCAmelCase_ ) assert generated_txt[0].strip() == tgt_text
83
_SCREAMING_SNAKE_CASE = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on _SCREAMING_SNAKE_CASE = {value: key for key, value in MORSE_CODE_DICT.items()} def snake_case ( snake_case__ :str) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper()) def snake_case ( snake_case__ :str) -> str: return "".join(REVERSE_DICT[char] for char in message.split()) def snake_case ( ) -> None: _A = """Morse code here!""" print(snake_case__) _A = encrypt(snake_case__) print(snake_case__) _A = decrypt(snake_case__) print(snake_case__) if __name__ == "__main__": main()
83
1
import itertools import json import os import unittest from transformers import AddedToken, RobertaTokenizer, RobertaTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = RobertaTokenizer lowerCamelCase :str = RobertaTokenizerFast lowerCamelCase :Dict = True lowerCamelCase :Optional[Any] = {'''cls_token''': '''<s>'''} def UpperCAmelCase ( self ) -> Union[str, Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _A = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] _A = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_ ) ) ) ) _A = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _A = {"""unk_token""": """<unk>"""} _A = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) _A = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowerCAmelCase_ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowerCAmelCase_ ) ) def UpperCAmelCase ( self , **lowerCAmelCase_ ) -> List[str]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def UpperCAmelCase ( self , **lowerCAmelCase_ ) -> List[str]: kwargs.update(self.special_tokens_map ) return RobertaTokenizerFast.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[str]: _A = """lower newer""" _A = """lower newer""" return input_text, output_text def UpperCAmelCase ( self ) -> Optional[Any]: _A = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) _A = """lower newer""" _A = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _A = tokenizer.tokenize(lowerCAmelCase_ ) # , add_prefix_space=True) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A = tokens + [tokenizer.unk_token] _A = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""" , add_special_tokens=lowerCAmelCase_ ) , [0, 3_14_14, 2_32, 3_28, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""" , add_special_tokens=lowerCAmelCase_ ) , [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2] , ) @slow def UpperCAmelCase ( self ) -> List[Any]: _A = self.tokenizer_class.from_pretrained("""roberta-base""" ) _A = tokenizer.encode("""sequence builders""" , add_special_tokens=lowerCAmelCase_ ) _A = tokenizer.encode("""multi-sequence build""" , add_special_tokens=lowerCAmelCase_ ) _A = tokenizer.encode( """sequence builders""" , add_special_tokens=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ ) _A = tokenizer.encode( """sequence builders""" , """multi-sequence build""" , add_special_tokens=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ ) _A = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) _A = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def UpperCAmelCase ( self ) -> Any: _A = self.get_tokenizer() _A = """Encode this sequence.""" _A = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments _A = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ ) _A = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ ) _A = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) _A = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) _A = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # Testing spaces after special tokens _A = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ )} ) # mask token has a left space _A = tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) _A = """Encode <mask> sequence""" _A = """Encode <mask>sequence""" _A = tokenizer.encode(lowerCAmelCase_ ) _A = encoded.index(lowerCAmelCase_ ) _A = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A = tokenizer.encode(lowerCAmelCase_ ) _A = encoded.index(lowerCAmelCase_ ) _A = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[str]: pass def UpperCAmelCase ( self ) -> str: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): _A = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) _A = self.tokenizer_class.from_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) _A = """A, <mask> AllenNLP sentence.""" _A = tokenizer_r.encode_plus(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ ) _A = tokenizer_p.encode_plus(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) _A = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) _A = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase_ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( lowerCAmelCase_ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def UpperCAmelCase ( self ) -> int: for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): _A = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) _A = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""] , lowerCAmelCase_ ) self.assertEqual(post_processor_state["""add_prefix_space"""] , lowerCAmelCase_ ) self.assertEqual(post_processor_state["""trim_offsets"""] , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[Any]: # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): _A = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` _A = F'''{text_of_1_token} {text_of_1_token}''' _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase_ ) + 1, len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase_ ) + 1, len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase_ ), len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase_ ), len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = F''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase_ ) + 1, 1 + len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase_ ), 1 + len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , ) _A = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase_ , use_fast=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , trim_offsets=lowerCAmelCase_ ) _A = tokenizer_r(lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowerCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase_ ), 1 + len(lowerCAmelCase_ ) + 1 + len(lowerCAmelCase_ )) , )
83
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _SCREAMING_SNAKE_CASE = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class a ( nn.Module ): """simple docstring""" lowerCamelCase :int lowerCamelCase :int lowerCamelCase :float = 0.0 lowerCamelCase :int = 1 lowerCamelCase :int = 1 lowerCamelCase :bool = True lowerCamelCase :bool = False lowerCamelCase :bool = False lowerCamelCase :bool = False lowerCamelCase :jnp.dtype = jnp.floataa def UpperCAmelCase ( self ) -> Union[str, Any]: _A = [] _A = [] for i in range(self.num_layers ): _A = self.in_channels if i == 0 else self.out_channels _A = FlaxResnetBlockaD( in_channels=lowerCAmelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCAmelCase_ ) _A = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCAmelCase_ ) _A = resnets _A = attentions if self.add_downsample: _A = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> List[Any]: _A = () for resnet, attn in zip(self.resnets , self.attentions ): _A = resnet(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) _A = attn(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) output_states += (hidden_states,) if self.add_downsample: _A = self.downsamplers_a(lowerCAmelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class a ( nn.Module ): """simple docstring""" lowerCamelCase :int lowerCamelCase :int lowerCamelCase :float = 0.0 lowerCamelCase :int = 1 lowerCamelCase :bool = True lowerCamelCase :jnp.dtype = jnp.floataa def UpperCAmelCase ( self ) -> Optional[int]: _A = [] for i in range(self.num_layers ): _A = self.in_channels if i == 0 else self.out_channels _A = FlaxResnetBlockaD( in_channels=lowerCAmelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCAmelCase_ ) _A = resnets if self.add_downsample: _A = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> Union[str, Any]: _A = () for resnet in self.resnets: _A = resnet(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) output_states += (hidden_states,) if self.add_downsample: _A = self.downsamplers_a(lowerCAmelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class a ( nn.Module ): """simple docstring""" lowerCamelCase :int lowerCamelCase :int lowerCamelCase :int lowerCamelCase :float = 0.0 lowerCamelCase :int = 1 lowerCamelCase :int = 1 lowerCamelCase :bool = True lowerCamelCase :bool = False lowerCamelCase :bool = False lowerCamelCase :bool = False lowerCamelCase :jnp.dtype = jnp.floataa def UpperCAmelCase ( self ) -> Optional[int]: _A = [] _A = [] for i in range(self.num_layers ): _A = self.in_channels if (i == self.num_layers - 1) else self.out_channels _A = self.prev_output_channel if i == 0 else self.out_channels _A = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCAmelCase_ ) _A = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCAmelCase_ ) _A = resnets _A = attentions if self.add_upsample: _A = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> Optional[int]: for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states _A = res_hidden_states_tuple[-1] _A = res_hidden_states_tuple[:-1] _A = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _A = resnet(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) _A = attn(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) if self.add_upsample: _A = self.upsamplers_a(lowerCAmelCase_ ) return hidden_states class a ( nn.Module ): """simple docstring""" lowerCamelCase :int lowerCamelCase :int lowerCamelCase :int lowerCamelCase :float = 0.0 lowerCamelCase :int = 1 lowerCamelCase :bool = True lowerCamelCase :jnp.dtype = jnp.floataa def UpperCAmelCase ( self ) -> Tuple: _A = [] for i in range(self.num_layers ): _A = self.in_channels if (i == self.num_layers - 1) else self.out_channels _A = self.prev_output_channel if i == 0 else self.out_channels _A = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCAmelCase_ ) _A = resnets if self.add_upsample: _A = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> Union[str, Any]: for resnet in self.resnets: # pop res hidden states _A = res_hidden_states_tuple[-1] _A = res_hidden_states_tuple[:-1] _A = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _A = resnet(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) if self.add_upsample: _A = self.upsamplers_a(lowerCAmelCase_ ) return hidden_states class a ( nn.Module ): """simple docstring""" lowerCamelCase :int lowerCamelCase :float = 0.0 lowerCamelCase :int = 1 lowerCamelCase :int = 1 lowerCamelCase :bool = False lowerCamelCase :bool = False lowerCamelCase :jnp.dtype = jnp.floataa def UpperCAmelCase ( self ) -> Optional[Any]: # there is always at least one resnet _A = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] _A = [] for _ in range(self.num_layers ): _A = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCAmelCase_ ) _A = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCAmelCase_ ) _A = resnets _A = attentions def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> List[str]: _A = self.resnets[0](lowerCAmelCase_ , lowerCAmelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): _A = attn(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) _A = resnet(lowerCAmelCase_ , lowerCAmelCase_ , deterministic=lowerCAmelCase_ ) return hidden_states
83
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''philschmid/bart-large-cnn-samsum''' lowerCamelCase :Tuple = ( '''This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ''' '''and returns a summary of the text.''' ) lowerCamelCase :List[Any] = '''summarizer''' lowerCamelCase :List[str] = AutoTokenizer lowerCamelCase :Dict = AutoModelForSeqaSeqLM lowerCamelCase :int = ['''text'''] lowerCamelCase :List[Any] = ['''text'''] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: return self.pre_processor(lowerCAmelCase_ , return_tensors="""pt""" , truncation=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: return self.model.generate(**lowerCAmelCase_ )[0] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: return self.pre_processor.decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ )
83
1
from torch import nn def snake_case ( snake_case__ :int) -> Dict: if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'''Unsupported activation function: {act_fn}''')
83
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _SCREAMING_SNAKE_CASE = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def snake_case ( snake_case__ :Union[str, Any]) -> Dict: _A = torch.load(snake_case__ , map_location="""cpu""") return sd def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any] , snake_case__ :int=rename_keys_prefix) -> Optional[Any]: _A = OrderedDict() _A = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _A = key for name_pair in rename_keys_prefix: _A = new_key.replace(name_pair[0] , name_pair[1]) _A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _A = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple) -> int: assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _A = """pretraining""" if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "nlvr" in checkpoint_path: _A = {"""visual_embedding_dim""": 1_024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} _A = """multichoice""" elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} _A = """vqa_advanced""" elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048, """num_labels""": 3_129} _A = """vqa""" elif "nlvr" in checkpoint_path: _A = { """visual_embedding_dim""": 1_024, """num_labels""": 2, } _A = """nlvr""" _A = VisualBertConfig(**snake_case__) # Load State Dict _A = load_state_dict(snake_case__) _A = get_new_dict(snake_case__ , snake_case__) if model_type == "pretraining": _A = VisualBertForPreTraining(snake_case__) elif model_type == "vqa": _A = VisualBertForQuestionAnswering(snake_case__) elif model_type == "nlvr": _A = VisualBertForVisualReasoning(snake_case__) elif model_type == "multichoice": _A = VisualBertForMultipleChoice(snake_case__) model.load_state_dict(snake_case__) # Save Checkpoints Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
83
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''speech_to_text_2''' lowerCamelCase :Dict = ['''past_key_values'''] lowerCamelCase :Dict = {'''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=2 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=10_24 , **lowerCAmelCase_ , ) -> List[Any]: _A = vocab_size _A = d_model _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = decoder_layerdrop _A = use_cache _A = decoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = max_target_positions super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , )
83
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def UpperCAmelCase ( self ) -> Optional[int]: _A = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCAmelCase_ ): self.assertDictEqual(lowerCAmelCase_ , example_records[i] ) def UpperCAmelCase ( self ) -> str: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) _A = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def UpperCAmelCase ( self ) -> Any: # checks what happens with missing columns _A = [{"""col_1""": 1}, {"""col_2""": """x"""}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def UpperCAmelCase ( self ) -> Tuple: # checks if the type can be inferred from the second record _A = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def UpperCAmelCase ( self ) -> Any: _A = Dataset.from_list([] ) self.assertEqual(len(lowerCAmelCase_ ) , 0 ) self.assertListEqual(dset.column_names , [] )
83
1
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch _SCREAMING_SNAKE_CASE = random.Random() def snake_case ( snake_case__ :int , snake_case__ :Tuple=1.0 , snake_case__ :Optional[Any]=None , snake_case__ :Any=None) -> Any: if rng is None: _A = global_rng _A = [] for batch_idx in range(shape[0]): values.append([]) for _ in range(shape[1]): values[-1].append(rng.random() * scale) return values @require_torch class a ( unittest.TestCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=7 , lowerCAmelCase_=4_00 , lowerCAmelCase_=20_00 , lowerCAmelCase_=1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=1_60_00 , lowerCAmelCase_=True , lowerCAmelCase_=80 , lowerCAmelCase_=16 , lowerCAmelCase_=64 , lowerCAmelCase_="hann_window" , lowerCAmelCase_=80 , lowerCAmelCase_=76_00 , lowerCAmelCase_=1E-10 , lowerCAmelCase_=True , ) -> int: _A = parent _A = batch_size _A = min_seq_length _A = max_seq_length _A = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _A = feature_size _A = padding_value _A = sampling_rate _A = do_normalize _A = num_mel_bins _A = hop_length _A = win_length _A = win_function _A = fmin _A = fmax _A = mel_floor _A = return_attention_mask def UpperCAmelCase ( self ) -> Tuple: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def UpperCAmelCase ( self , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Optional[Any]: def _flatten(lowerCAmelCase_ ): return list(itertools.chain(*lowerCAmelCase_ ) ) if equal_length: _A = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _A = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _A = [np.asarray(lowerCAmelCase_ ) for x in speech_inputs] return speech_inputs def UpperCAmelCase ( self , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Union[str, Any]: if equal_length: _A = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _A = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _A = [np.asarray(lowerCAmelCase_ ) for x in speech_inputs] return speech_inputs @require_torch class a ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = SpeechTaFeatureExtractor def UpperCAmelCase ( self ) -> Union[str, Any]: _A = SpeechTaFeatureExtractionTester(self ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: self.assertTrue(np.all(np.mean(lowerCAmelCase_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase_ , axis=0 ) - 1 ) < 1E-3 ) ) def UpperCAmelCase ( self ) -> Any: # Tests that all call wrap to encode_plus and batch_encode_plus _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = [np.asarray(lowerCAmelCase_ ) for speech_input in speech_inputs] # Test not batched input _A = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values _A = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test batched _A = feat_extract(lowerCAmelCase_ , return_tensors="""np""" ).input_values _A = feat_extract(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) def UpperCAmelCase ( self ) -> Any: _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = ["""longest""", """max_length""", """do_not_pad"""] _A = [None, 16_00, None] for max_length, padding in zip(lowerCAmelCase_ , lowerCAmelCase_ ): _A = feat_extract(lowerCAmelCase_ , padding=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors="""np""" ) _A = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self.assertTrue(input_values[0][8_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self.assertTrue(input_values[0][10_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _A = range(8_00 , 14_00 , 2_00 ) _A = [floats_list((1, x) )[0] for x in lengths] _A = ["""longest""", """max_length""", """do_not_pad"""] _A = [None, 16_00, None] for max_length, padding in zip(lowerCAmelCase_ , lowerCAmelCase_ ): _A = feat_extract(lowerCAmelCase_ , max_length=lowerCAmelCase_ , padding=lowerCAmelCase_ ) _A = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def UpperCAmelCase ( self ) -> Tuple: _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=10_00 , padding="""max_length""" , return_tensors="""np""" ) _A = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCAmelCase ( self ) -> List[Any]: _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=10_00 , padding="""longest""" , return_tensors="""np""" ) _A = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00) ) _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=20_00 , padding="""longest""" , return_tensors="""np""" ) _A = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00) ) def UpperCAmelCase ( self ) -> int: _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _A = np.random.rand(1_00 ).astype(np.floataa ) _A = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _A = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _A = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def UpperCAmelCase ( self ) -> str: # Tests that all call wrap to encode_plus and batch_encode_plus _A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _A = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] _A = [np.asarray(lowerCAmelCase_ ) for speech_input in speech_inputs] # Test feature size _A = feature_extractor(audio_target=lowerCAmelCase_ , padding=lowerCAmelCase_ , return_tensors="""np""" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input _A = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_values _A = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test batched _A = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values _A = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. _A = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] _A = np.asarray(lowerCAmelCase_ ) _A = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values _A = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) def UpperCAmelCase ( self ) -> int: _A = self.feat_extract_tester.prepare_inputs_for_target() _A = self.feature_extraction_class(**self.feat_extract_dict ) _A = feat_extract.model_input_names[0] _A = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) for x, y in zip(lowerCAmelCase_ , processed_features[input_name] ) ) ) _A = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase_ ) _A = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) _A = processed_features[input_name] if len(batch_features_input.shape ) < 3: _A = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def UpperCAmelCase ( self ) -> Optional[Any]: _A = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase_ ) _A = self.feature_extraction_class(**self.feat_extract_dict ) _A = feat_extract.model_input_names[0] _A = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) _A = processed_features[input_name] if len(batch_features_input.shape ) < 3: _A = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def UpperCAmelCase ( self ) -> List[str]: _A = self.feature_extraction_class(**self.feat_extract_dict ) _A = self.feat_extract_tester.prepare_inputs_for_target() _A = feat_extract.model_input_names[0] _A = BatchFeature({input_name: speech_inputs} ) _A = feat_extract.num_mel_bins # hack! _A = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""np""" )[input_name] _A = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = self.feat_extract_dict _A = True _A = self.feature_extraction_class(**lowerCAmelCase_ ) _A = self.feat_extract_tester.prepare_inputs_for_target() _A = [len(lowerCAmelCase_ ) for x in speech_inputs] _A = feat_extract.model_input_names[0] _A = BatchFeature({input_name: speech_inputs} ) _A = feat_extract.num_mel_bins # hack! _A = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , lowerCAmelCase_ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = self.feat_extract_dict _A = True _A = self.feature_extraction_class(**lowerCAmelCase_ ) _A = self.feat_extract_tester.prepare_inputs_for_target() _A = [len(lowerCAmelCase_ ) for x in speech_inputs] _A = feat_extract.model_input_names[0] _A = BatchFeature({input_name: speech_inputs} ) _A = min(lowerCAmelCase_ ) _A = feat_extract.num_mel_bins # hack! _A = feat_extract.pad( lowerCAmelCase_ , padding="""max_length""" , max_length=lowerCAmelCase_ , truncation=lowerCAmelCase_ , return_tensors="""np""" ) self.assertIn("""attention_mask""" , lowerCAmelCase_ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: from datasets import load_dataset _A = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech _A = ds.sort("""id""" ).select(range(lowerCAmelCase_ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def UpperCAmelCase ( self ) -> Any: # fmt: off _A = torch.tensor( [2.3804E-03, 2.0752E-03, 1.9836E-03, 2.1057E-03, 1.6174E-03, 3.0518E-04, 9.1553E-05, 3.3569E-04, 9.7656E-04, 1.8311E-03, 2.0142E-03, 2.1057E-03, 1.7395E-03, 4.5776E-04, -3.9673E-04, 4.5776E-04, 1.0071E-03, 9.1553E-05, 4.8828E-04, 1.1597E-03, 7.3242E-04, 9.4604E-04, 1.8005E-03, 1.8311E-03, 8.8501E-04, 4.2725E-04, 4.8828E-04, 7.3242E-04, 1.0986E-03, 2.1057E-03] ) # fmt: on _A = self._load_datasamples(1 ) _A = SpeechTaFeatureExtractor() _A = feature_extractor(lowerCAmelCase_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 9_36_80) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase_ , atol=1E-6 ) ) def UpperCAmelCase ( self ) -> List[Any]: # fmt: off _A = torch.tensor( [-2.6870, -3.0104, -3.1356, -3.5352, -3.0044, -3.0353, -3.4719, -3.6777, -3.1520, -2.9435, -2.6553, -2.8795, -2.9944, -2.5921, -3.0279, -3.0386, -3.0864, -3.1291, -3.2353, -2.7444, -2.6831, -2.7287, -3.1761, -3.1571, -3.2726, -3.0582, -3.1007, -3.4533, -3.4695, -3.0998] ) # fmt: on _A = self._load_datasamples(1 ) _A = SpeechTaFeatureExtractor() _A = feature_extractor(audio_target=lowerCAmelCase_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 3_66, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase_ , atol=1E-4 ) )
83
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case__ , snake_case__))) _A = [float(snake_case__) for n in range(limit + 1)] for p in primes: for n in range(snake_case__ , limit + 1 , snake_case__): phi[n] *= 1 - 1 / p return int(sum(phi[2:])) if __name__ == "__main__": print(F'''{solution() = }''')
83
1
from __future__ import annotations from fractions import Fraction def snake_case ( snake_case__ :int , snake_case__ :int) -> bool: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def snake_case ( snake_case__ :int) -> list[str]: _A = [] _A = 11 _A = int("""1""" + """0""" * digit_len) for num in range(snake_case__ , snake_case__): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(snake_case__ , snake_case__): solutions.append(F'''{num}/{den}''') den += 1 num += 1 _A = 10 return solutions def snake_case ( snake_case__ :int = 2) -> int: _A = 1.0 for fraction in fraction_list(snake_case__): _A = Fraction(snake_case__) result *= frac.denominator / frac.numerator return int(snake_case__) if __name__ == "__main__": print(solution())
83
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=32 , lowerCAmelCase_=5 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_="None" , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=None , ) -> Union[str, Any]: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = num_choices _A = relative_attention _A = position_biased_input _A = pos_att_type _A = scope def UpperCAmelCase ( self ) -> Dict: _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self ) -> Optional[int]: return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Any: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: _A = DebertaVaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = DebertaVaForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: _A = self.num_labels _A = DebertaVaForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = self.num_labels _A = DebertaVaForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _A = DebertaVaForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = DebertaVaForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) lowerCamelCase :str = ( { '''feature-extraction''': DebertaVaModel, '''fill-mask''': DebertaVaForMaskedLM, '''question-answering''': DebertaVaForQuestionAnswering, '''text-classification''': DebertaVaForSequenceClassification, '''token-classification''': DebertaVaForTokenClassification, '''zero-shot''': DebertaVaForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase :str = True lowerCamelCase :Union[str, Any] = False lowerCamelCase :Optional[int] = False lowerCamelCase :List[str] = False lowerCamelCase :str = False def UpperCAmelCase ( self ) -> Optional[int]: _A = DebertaVaModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> List[str]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> List[str]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> Any: for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = DebertaVaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason="""Model not available yet""" ) def UpperCAmelCase ( self ) -> int: pass @slow def UpperCAmelCase ( self ) -> Optional[Any]: _A = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _A = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) _A = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] # compare the actual values for a slice. _A = torch.tensor( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
83
1
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( """files""" , [ ["""full:README.md""", """dataset_infos.json"""], ["""empty:README.md""", """dataset_infos.json"""], ["""dataset_infos.json"""], ["""full:README.md"""], ] , ) def snake_case ( snake_case__ :Optional[int] , snake_case__ :int) -> List[Any]: _A = tmp_path_factory.mktemp("""dset_infos_dir""") if "full:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""") as f: f.write("""---\ndataset_info:\n dataset_size: 42\n---""") if "empty:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""") as f: f.write("""""") # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / """dataset_infos.json""" , """w""") as f: f.write("""{\"default\": {\"dataset_size\": 42}}""") _A = DatasetInfosDict.from_directory(snake_case__) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( """dataset_info""" , [ DatasetInfo(), DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""")}) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ), ] , ) def snake_case ( snake_case__ :Optional[Any] , snake_case__ :DatasetInfo) -> Any: _A = str(snake_case__) dataset_info.write_to_directory(snake_case__) _A = DatasetInfo.from_directory(snake_case__) assert dataset_info == reloaded assert os.path.exists(os.path.join(snake_case__ , """dataset_info.json""")) def snake_case ( ) -> str: _A = DatasetInfo( description="""foo""" , citation="""bar""" , homepage="""https://foo.bar""" , license="""CC0""" , features=Features({"""a""": Value("""int32""")}) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train""", """num_examples""": 42}] , download_checksums={} , download_size=1_337 , post_processing_size=442 , dataset_size=1_234 , size_in_bytes=1_337 + 442 + 1_234 , ) _A = dataset_info._to_yaml_dict() assert sorted(snake_case__) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str)) _A = yaml.safe_dump(snake_case__) _A = yaml.safe_load(snake_case__) assert dataset_info_yaml_dict == reloaded def snake_case ( ) -> Optional[Any]: _A = DatasetInfo() _A = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( """dataset_infos_dict""" , [ DatasetInfosDict(), DatasetInfosDict({"""default""": DatasetInfo()}), DatasetInfosDict({"""my_config_name""": DatasetInfo()}), DatasetInfosDict( { """default""": DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""")}) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ) }), DatasetInfosDict( { """v1""": DatasetInfo(dataset_size=42), """v2""": DatasetInfo(dataset_size=1_337), }), ] , ) def snake_case ( snake_case__ :Tuple , snake_case__ :DatasetInfosDict) -> List[str]: _A = str(snake_case__) dataset_infos_dict.write_to_directory(snake_case__) _A = DatasetInfosDict.from_directory(snake_case__) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): _A = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml _A = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict()) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(snake_case__ , """README.md"""))
83
def snake_case ( snake_case__ :int , snake_case__ :int) -> int: return int(input_a == input_a == 0) def snake_case ( ) -> None: print("""Truth Table of NOR Gate:""") print("""| Input 1 | Input 2 | Output |""") print(F'''| 0 | 0 | {nor_gate(0 , 0)} |''') print(F'''| 0 | 1 | {nor_gate(0 , 1)} |''') print(F'''| 1 | 0 | {nor_gate(1 , 0)} |''') print(F'''| 1 | 1 | {nor_gate(1 , 1)} |''') if __name__ == "__main__": import doctest doctest.testmod() main()
83
1
def snake_case ( snake_case__ :int , snake_case__ :list) -> Any: _enforce_args(snake_case__ , snake_case__) if n == 0: return 0 _A = float("""-inf""") for i in range(1 , n + 1): _A = max( snake_case__ , prices[i - 1] + naive_cut_rod_recursive(n - i , snake_case__)) return max_revue def snake_case ( snake_case__ :int , snake_case__ :list) -> Optional[Any]: _enforce_args(snake_case__ , snake_case__) _A = [float("""-inf""") for _ in range(n + 1)] return _top_down_cut_rod_recursive(snake_case__ , snake_case__ , snake_case__) def snake_case ( snake_case__ :int , snake_case__ :list , snake_case__ :list) -> Optional[Any]: if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: _A = float("""-inf""") for i in range(1 , n + 1): _A = max( snake_case__ , prices[i - 1] + _top_down_cut_rod_recursive(n - i , snake_case__ , snake_case__) , ) _A = max_revenue return max_rev[n] def snake_case ( snake_case__ :int , snake_case__ :list) -> Optional[Any]: _enforce_args(snake_case__ , snake_case__) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. _A = [float("""-inf""") for _ in range(n + 1)] _A = 0 for i in range(1 , n + 1): _A = max_rev[i] for j in range(1 , i + 1): _A = max(snake_case__ , prices[j - 1] + max_rev[i - j]) _A = max_revenue_i return max_rev[n] def snake_case ( snake_case__ :int , snake_case__ :list) -> Union[str, Any]: if n < 0: _A = F'''n must be greater than or equal to 0. Got n = {n}''' raise ValueError(snake_case__) if n > len(snake_case__): _A = ( """Each integral piece of rod must have a corresponding price. """ F'''Got n = {n} but length of prices = {len(snake_case__)}''' ) raise ValueError(snake_case__) def snake_case ( ) -> List[str]: _A = [6, 10, 12, 15, 20, 23] _A = len(snake_case__) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. _A = 36 _A = top_down_cut_rod(snake_case__ , snake_case__) _A = bottom_up_cut_rod(snake_case__ , snake_case__) _A = naive_cut_rod_recursive(snake_case__ , snake_case__) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
83
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=sys.maxsize ) -> str: _A = """bilinear""" _A = max_size _A = short_edge_length def __call__( self , lowerCAmelCase_ ) -> Optional[Any]: _A = [] for img in imgs: _A , _A = img.shape[:2] # later: provide list and randomly choose index for resize _A = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img _A = size * 1.0 / min(lowerCAmelCase_ , lowerCAmelCase_ ) if h < w: _A , _A = size, scale * w else: _A , _A = scale * h, size if max(lowerCAmelCase_ , lowerCAmelCase_ ) > self.max_size: _A = self.max_size * 1.0 / max(lowerCAmelCase_ , lowerCAmelCase_ ) _A = newh * scale _A = neww * scale _A = int(neww + 0.5 ) _A = int(newh + 0.5 ) if img.dtype == np.uinta: _A = Image.fromarray(lowerCAmelCase_ ) _A = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) _A = np.asarray(lowerCAmelCase_ ) else: _A = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw _A = nn.functional.interpolate( lowerCAmelCase_ , (newh, neww) , mode=self.interp_method , align_corners=lowerCAmelCase_ ).squeeze(0 ) img_augs.append(lowerCAmelCase_ ) return img_augs class a : """simple docstring""" def __init__( self , lowerCAmelCase_ ) -> List[Any]: _A = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) _A = cfg.INPUT.FORMAT _A = cfg.SIZE_DIVISIBILITY _A = cfg.PAD_VALUE _A = cfg.INPUT.MAX_SIZE_TEST _A = cfg.MODEL.DEVICE _A = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = lambda lowerCAmelCase_ : (x - self.pixel_mean) / self.pixel_std def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = tuple(max(lowerCAmelCase_ ) for s in zip(*[img.shape for img in images] ) ) _A = [im.shape[-2:] for im in images] _A = [ nn.functional.pad( lowerCAmelCase_ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] return torch.stack(lowerCAmelCase_ ), torch.tensor(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_=False ) -> int: with torch.no_grad(): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = [images] if single_image: assert len(lowerCAmelCase_ ) == 1 for i in range(len(lowerCAmelCase_ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(lowerCAmelCase_ , images.pop(lowerCAmelCase_ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( lowerCAmelCase_ , torch.as_tensor(img_tensorize(images.pop(lowerCAmelCase_ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge _A = torch.tensor([im.shape[:2] for im in images] ) _A = self.aug(lowerCAmelCase_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic _A = [self.normalizer(lowerCAmelCase_ ) for x in images] # now pad them to do the following operations _A , _A = self.pad(lowerCAmelCase_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad _A = torch.true_divide(lowerCAmelCase_ , lowerCAmelCase_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[Any]) -> Tuple: boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def snake_case ( snake_case__ :Optional[int] , snake_case__ :Tuple[int, int]) -> Optional[Any]: assert torch.isfinite(snake_case__).all(), "Box tensor contains infinite or NaN!" _A , _A = box_size tensor[:, 0].clamp_(min=0 , max=snake_case__) tensor[:, 1].clamp_(min=0 , max=snake_case__) tensor[:, 2].clamp_(min=0 , max=snake_case__) tensor[:, 3].clamp_(min=0 , max=snake_case__)
83
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = '''gpt_neox''' def __init__( self , lowerCAmelCase_=5_04_32 , lowerCAmelCase_=61_44 , lowerCAmelCase_=44 , lowerCAmelCase_=64 , lowerCAmelCase_=2_45_76 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.25 , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.1 , lowerCAmelCase_=20_48 , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-5 , lowerCAmelCase_=True , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_=None , **lowerCAmelCase_ , ) -> Dict: super().__init__(bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) _A = vocab_size _A = max_position_embeddings _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = rotary_pct _A = rotary_emb_base _A = attention_dropout _A = hidden_dropout _A = classifier_dropout _A = initializer_range _A = layer_norm_eps _A = use_cache _A = tie_word_embeddings _A = use_parallel_residual _A = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( """The hidden size is not divisble by the number of attention heads! Make sure to update them!""" ) def UpperCAmelCase ( self ) -> Dict: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , lowerCAmelCase_ ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ F'''got {self.rope_scaling}''' ) _A = self.rope_scaling.get("""type""" , lowerCAmelCase_ ) _A = self.rope_scaling.get("""factor""" , lowerCAmelCase_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or rope_scaling_factor <= 1.0: raise ValueError(F'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
83
from collections import defaultdict def snake_case ( snake_case__ :int) -> int: _A = 1 _A = True for v in tree[start]: if v not in visited: ret += dfs(snake_case__) if ret % 2 == 0: cuts.append(snake_case__) return ret def snake_case ( ) -> Any: dfs(1) if __name__ == "__main__": _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10, 9 _SCREAMING_SNAKE_CASE = defaultdict(list) _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
83
1
def snake_case ( snake_case__ :float) -> float: return 10 - x * x def snake_case ( snake_case__ :float , snake_case__ :float) -> float: # Bolzano theory in order to find if there is a root between a and b if equation(snake_case__) * equation(snake_case__) >= 0: raise ValueError("""Wrong space!""") _A = a while (b - a) >= 0.01: # Find middle point _A = (a + b) / 2 # Check if middle point is root if equation(snake_case__) == 0.0: break # Decide the side to repeat the steps if equation(snake_case__) * equation(snake_case__) < 0: _A = c else: _A = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
83
import heapq def snake_case ( snake_case__ :dict) -> set[int]: _A = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(snake_case__ , [-1 * len(snake_case__), (key, value)]) # chosen_vertices = set of chosen vertices _A = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _A = heapq.heappop(snake_case__)[1][0] chosen_vertices.add(snake_case__) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _A = elem[1][1].index(snake_case__) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(snake_case__) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(F'''Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}''')
83
1
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _A = FlaxDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ ) _A = [t[-1] for t in os.walk(os.path.join(lowerCAmelCase_ , os.listdir(lowerCAmelCase_ )[0] , """snapshots""" ) )] _A = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(""".bin""" ) for f in files ) @slow @require_flax class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Tuple: _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=lowerCAmelCase_ ) _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.random.PRNGKey(0 ) _A = 4 _A = jax.device_count() _A = num_samples * [prompt] _A = pipeline.prepare_inputs(lowerCAmelCase_ ) # shard inputs and rng _A = replicate(lowerCAmelCase_ ) _A = jax.random.split(lowerCAmelCase_ , lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.151_4745 ) < 1E-3 assert np.abs(np.abs(lowerCAmelCase_ , dtype=np.floataa ).sum() - 4_9947.875 ) < 5E-1 _A = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowerCAmelCase_ ) == num_samples def UpperCAmelCase ( self ) -> List[Any]: _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""flax""" , safety_checker=lowerCAmelCase_ ) _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.random.PRNGKey(0 ) _A = 50 _A = jax.device_count() _A = num_samples * [prompt] _A = pipeline.prepare_inputs(lowerCAmelCase_ ) # shard inputs and rng _A = replicate(lowerCAmelCase_ ) _A = jax.random.split(lowerCAmelCase_ , lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 5_12, 5_12, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0565_2401) ) < 1E-3 assert np.abs((np.abs(lowerCAmelCase_ , dtype=np.floataa ).sum() - 238_3808.2) ) < 5E-1 def UpperCAmelCase ( self ) -> Optional[Any]: _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=lowerCAmelCase_ ) _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.random.PRNGKey(0 ) _A = 50 _A = jax.device_count() _A = num_samples * [prompt] _A = pipeline.prepare_inputs(lowerCAmelCase_ ) # shard inputs and rng _A = replicate(lowerCAmelCase_ ) _A = jax.random.split(lowerCAmelCase_ , lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 5_12, 5_12, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(lowerCAmelCase_ , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def UpperCAmelCase ( self ) -> Optional[Any]: _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa ) _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.random.PRNGKey(0 ) _A = 50 _A = jax.device_count() _A = num_samples * [prompt] _A = pipeline.prepare_inputs(lowerCAmelCase_ ) # shard inputs and rng _A = replicate(lowerCAmelCase_ ) _A = jax.random.split(lowerCAmelCase_ , lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 5_12, 5_12, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(lowerCAmelCase_ , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def UpperCAmelCase ( self ) -> Union[str, Any]: _A = FlaxDDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , set_alpha_to_one=lowerCAmelCase_ , steps_offset=1 , ) _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , scheduler=lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) _A = scheduler.create_state() _A = scheduler_state _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.random.PRNGKey(0 ) _A = 50 _A = jax.device_count() _A = num_samples * [prompt] _A = pipeline.prepare_inputs(lowerCAmelCase_ ) # shard inputs and rng _A = replicate(lowerCAmelCase_ ) _A = jax.random.split(lowerCAmelCase_ , lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 5_12, 5_12, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_4504_3945) ) < 1E-3 assert np.abs((np.abs(lowerCAmelCase_ , dtype=np.floataa ).sum() - 234_7693.5) ) < 5E-1 def UpperCAmelCase ( self ) -> str: _A = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) _A = jax.device_count() _A = num_samples * [prompt] _A = jax.random.split(jax.random.PRNGKey(0 ) , lowerCAmelCase_ ) _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=lowerCAmelCase_ , ) _A = replicate(lowerCAmelCase_ ) _A = pipeline.prepare_inputs(lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images.shape == (num_samples, 1, 5_12, 5_12, 3) _A = images[2, 0, 2_56, 10:17, 1] # With memory efficient attention _A , _A = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=lowerCAmelCase_ , use_memory_efficient_attention=lowerCAmelCase_ , ) _A = replicate(lowerCAmelCase_ ) _A = pipeline.prepare_inputs(lowerCAmelCase_ ) _A = shard(lowerCAmelCase_ ) _A = pipeline(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , jit=lowerCAmelCase_ ).images assert images_eff.shape == (num_samples, 1, 5_12, 5_12, 3) _A = images[2, 0, 2_56, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
83
import math import unittest def snake_case ( snake_case__ :int) -> bool: assert isinstance(snake_case__ , snake_case__) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def UpperCAmelCase ( self ) -> Dict: with self.assertRaises(lowerCAmelCase_ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , """Zero doesn't have any positive factors, primes must have exactly two.""" , ) self.assertFalse( is_prime(1 ) , """One only has 1 positive factor, primes must have exactly two.""" , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
83
1
_SCREAMING_SNAKE_CASE = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on _SCREAMING_SNAKE_CASE = {value: key for key, value in MORSE_CODE_DICT.items()} def snake_case ( snake_case__ :str) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper()) def snake_case ( snake_case__ :str) -> str: return "".join(REVERSE_DICT[char] for char in message.split()) def snake_case ( ) -> None: _A = """Morse code here!""" print(snake_case__) _A = encrypt(snake_case__) print(snake_case__) _A = decrypt(snake_case__) print(snake_case__) if __name__ == "__main__": main()
83
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'naver-clova-ix/donut-base': 'https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json', # See all Donut models at https://huggingface.co/models?filter=donut-swin } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Optional[int] = '''donut-swin''' lowerCamelCase :str = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self , lowerCAmelCase_=2_24 , lowerCAmelCase_=4 , lowerCAmelCase_=3 , lowerCAmelCase_=96 , lowerCAmelCase_=[2, 2, 6, 2] , lowerCAmelCase_=[3, 6, 12, 24] , lowerCAmelCase_=7 , lowerCAmelCase_=4.0 , lowerCAmelCase_=True , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.1 , lowerCAmelCase_="gelu" , lowerCAmelCase_=False , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-5 , **lowerCAmelCase_ , ) -> Tuple: super().__init__(**lowerCAmelCase_ ) _A = image_size _A = patch_size _A = num_channels _A = embed_dim _A = depths _A = len(lowerCAmelCase_ ) _A = num_heads _A = window_size _A = mlp_ratio _A = qkv_bias _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = drop_path_rate _A = hidden_act _A = use_absolute_embeddings _A = layer_norm_eps _A = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _A = int(embed_dim * 2 ** (len(lowerCAmelCase_ ) - 1) )
83
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(__lowerCAmelCase ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Optional[Any]: super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) self.check_model_type(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , **lowerCAmelCase_ ) -> Tuple: _A , _A = {}, {} if padding is not None: _A = padding if truncation is not None: _A = truncation if top_k is not None: _A = top_k return preprocess_params, {}, postprocess_params def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , **lowerCAmelCase_ ) -> Union[str, Any]: if isinstance(lowerCAmelCase_ , (Image.Image, str) ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = {"""image""": image, """question""": question} else: _A = image _A = super().__call__(lowerCAmelCase_ , **lowerCAmelCase_ ) return results def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Any: _A = load_image(inputs["""image"""] ) _A = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ ) _A = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase_ ) return model_inputs def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = self.model(**lowerCAmelCase_ ) return model_outputs def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=5 ) -> Union[str, Any]: if top_k > self.model.config.num_labels: _A = self.model.config.num_labels if self.framework == "pt": _A = model_outputs.logits.sigmoid()[0] _A , _A = probs.topk(lowerCAmelCase_ ) else: raise ValueError(F'''Unsupported framework: {self.framework}''' ) _A = scores.tolist() _A = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowerCAmelCase_ , lowerCAmelCase_ )]
83
1
import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def snake_case ( snake_case__ :int) -> Union[str, Any]: random.seed(snake_case__) np.random.seed(snake_case__) torch.manual_seed(snake_case__) torch.cuda.manual_seed_all(snake_case__) # ^^ safe to call this function even if cuda is not available class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ = 0.9999 , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 0 , lowerCAmelCase_ = False , lowerCAmelCase_ = 1.0 , lowerCAmelCase_ = 2 / 3 , lowerCAmelCase_ = None , lowerCAmelCase_ = None , **lowerCAmelCase_ , ) -> Optional[Any]: if isinstance(lowerCAmelCase_ , torch.nn.Module ): _A = ( """Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. """ """Please pass the parameters of the module instead.""" ) deprecate( """passing a `torch.nn.Module` to `ExponentialMovingAverage`""" , """1.0.0""" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ , ) _A = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _A = True if kwargs.get("""max_value""" , lowerCAmelCase_ ) is not None: _A = """The `max_value` argument is deprecated. Please use `decay` instead.""" deprecate("""max_value""" , """1.0.0""" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) _A = kwargs["""max_value"""] if kwargs.get("""min_value""" , lowerCAmelCase_ ) is not None: _A = """The `min_value` argument is deprecated. Please use `min_decay` instead.""" deprecate("""min_value""" , """1.0.0""" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) _A = kwargs["""min_value"""] _A = list(lowerCAmelCase_ ) _A = [p.clone().detach() for p in parameters] if kwargs.get("""device""" , lowerCAmelCase_ ) is not None: _A = """The `device` argument is deprecated. Please use `to` instead.""" deprecate("""device""" , """1.0.0""" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) self.to(device=kwargs["""device"""] ) _A = None _A = decay _A = min_decay _A = update_after_step _A = use_ema_warmup _A = inv_gamma _A = power _A = 0 _A = None # set in `step()` _A = model_cls _A = model_config @classmethod def UpperCAmelCase ( cls , lowerCAmelCase_ , lowerCAmelCase_ ) -> "EMAModel": _A , _A = model_cls.load_config(lowerCAmelCase_ , return_unused_kwargs=lowerCAmelCase_ ) _A = model_cls.from_pretrained(lowerCAmelCase_ ) _A = cls(model.parameters() , model_cls=lowerCAmelCase_ , model_config=model.config ) ema_model.load_state_dict(lowerCAmelCase_ ) return ema_model def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Optional[Any]: if self.model_cls is None: raise ValueError("""`save_pretrained` can only be used if `model_cls` was defined at __init__.""" ) if self.model_config is None: raise ValueError("""`save_pretrained` can only be used if `model_config` was defined at __init__.""" ) _A = self.model_cls.from_config(self.model_config ) _A = self.state_dict() state_dict.pop("""shadow_params""" , lowerCAmelCase_ ) model.register_to_config(**lowerCAmelCase_ ) self.copy_to(model.parameters() ) model.save_pretrained(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> float: _A = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _A = 1 - (1 + step / self.inv_gamma) ** -self.power else: _A = (1 + step) / (10 + step) _A = min(lowerCAmelCase_ , self.decay ) # make sure decay is not smaller than min_decay _A = max(lowerCAmelCase_ , self.min_decay ) return cur_decay_value @torch.no_grad() def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Optional[int]: if isinstance(lowerCAmelCase_ , torch.nn.Module ): _A = ( """Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. """ """Please pass the parameters of the module instead.""" ) deprecate( """passing a `torch.nn.Module` to `ExponentialMovingAverage.step`""" , """1.0.0""" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ , ) _A = parameters.parameters() _A = list(lowerCAmelCase_ ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _A = self.get_decay(self.optimization_step ) _A = decay _A = 1 - decay _A = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , lowerCAmelCase_ ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _A = deepspeed.zero.GatheredParameters(lowerCAmelCase_ , modifier_rank=lowerCAmelCase_ ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = list(lowerCAmelCase_ ) for s_param, param in zip(self.shadow_params , lowerCAmelCase_ ): param.data.copy_(s_param.to(param.device ).data ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None ) -> None: _A = [ p.to(device=lowerCAmelCase_ , dtype=lowerCAmelCase_ ) if p.is_floating_point() else p.to(device=lowerCAmelCase_ ) for p in self.shadow_params ] def UpperCAmelCase ( self ) -> dict: return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = [param.detach().cpu().clone() for param in parameters] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: if self.temp_stored_params is None: raise RuntimeError("""This ExponentialMovingAverage has no `store()`ed weights """ """to `restore()`""" ) for c_param, param in zip(self.temp_stored_params , lowerCAmelCase_ ): param.data.copy_(c_param.data ) # Better memory-wise. _A = None def UpperCAmelCase ( self , lowerCAmelCase_ ) -> None: _A = copy.deepcopy(lowerCAmelCase_ ) _A = state_dict.get("""decay""" , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError("""Decay must be between 0 and 1""" ) _A = state_dict.get("""min_decay""" , self.min_decay ) if not isinstance(self.min_decay , lowerCAmelCase_ ): raise ValueError("""Invalid min_decay""" ) _A = state_dict.get("""optimization_step""" , self.optimization_step ) if not isinstance(self.optimization_step , lowerCAmelCase_ ): raise ValueError("""Invalid optimization_step""" ) _A = state_dict.get("""update_after_step""" , self.update_after_step ) if not isinstance(self.update_after_step , lowerCAmelCase_ ): raise ValueError("""Invalid update_after_step""" ) _A = state_dict.get("""use_ema_warmup""" , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , lowerCAmelCase_ ): raise ValueError("""Invalid use_ema_warmup""" ) _A = state_dict.get("""inv_gamma""" , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError("""Invalid inv_gamma""" ) _A = state_dict.get("""power""" , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError("""Invalid power""" ) _A = state_dict.get("""shadow_params""" , lowerCAmelCase_ ) if shadow_params is not None: _A = shadow_params if not isinstance(self.shadow_params , lowerCAmelCase_ ): raise ValueError("""shadow_params must be a list""" ) if not all(isinstance(lowerCAmelCase_ , torch.Tensor ) for p in self.shadow_params ): raise ValueError("""shadow_params must all be Tensors""" )
83
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :str , snake_case__ :PreTrainedTokenizer , snake_case__ :int , snake_case__ :Optional[int] = None , ) -> Optional[int]: _A = {} if train_file is not None: _A = [train_file] if eval_file is not None: _A = [eval_file] if test_file is not None: _A = [test_file] _A = datasets.load_dataset("""csv""" , data_files=snake_case__) _A = list(ds[list(files.keys())[0]].features.keys()) _A = features_name.pop(snake_case__) _A = list(set(ds[list(files.keys())[0]][label_name])) _A = {label: i for i, label in enumerate(snake_case__)} _A = tokenizer.model_input_names _A = {} if len(snake_case__) == 1: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( example[features_name[0]] , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""") , batched=snake_case__ , ) elif len(snake_case__) == 2: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""" , ) , batched=snake_case__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _A = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _A = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _A = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST]))) return train_ds, val_ds, test_ds, labelaid _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class a : """simple docstring""" lowerCamelCase :int = field(metadata={'''help''': '''Which column contains the label'''} ) lowerCamelCase :str = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the training file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the development file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the test file'''} ) lowerCamelCase :int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCamelCase :bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) @dataclass class a : """simple docstring""" lowerCamelCase :str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) lowerCamelCase :bool = field(default=__lowerCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) def snake_case ( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments)) _A , _A , _A = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir) and os.listdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""") # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info( F'''n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1)}, ''' F'''16-bits training: {training_args.fpaa}''') logger.info(F'''Training/evaluation parameters {training_args}''') # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _A = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _A , _A , _A , _A = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=snake_case__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _A = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(snake_case__) , labelaid=snake_case__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="""text-classification""" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _A = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(""".bin""" in model_args.model_name_or_path) , config=snake_case__ , cache_dir=model_args.cache_dir , ) def compute_metrics(snake_case__ :EvalPrediction) -> Dict: _A = np.argmax(p.predictions , axis=1) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _A = TFTrainer( model=snake_case__ , args=snake_case__ , train_dataset=snake_case__ , eval_dataset=snake_case__ , compute_metrics=snake_case__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir) # Evaluation _A = {} if training_args.do_eval: logger.info("""*** Evaluate ***""") _A = trainer.evaluate() _A = os.path.join(training_args.output_dir , """eval_results.txt""") with open(snake_case__ , """w""") as writer: logger.info("""***** Eval results *****""") for key, value in result.items(): logger.info(F''' {key} = {value}''') writer.write(F'''{key} = {value}\n''') results.update(snake_case__) return results if __name__ == "__main__": main()
83
1
from __future__ import annotations _SCREAMING_SNAKE_CASE = tuple[int, int, int] _SCREAMING_SNAKE_CASE = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase _SCREAMING_SNAKE_CASE = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' # -------------------------- default selection -------------------------- # rotors -------------------------- _SCREAMING_SNAKE_CASE = 'EGZWVONAHDCLFQMSIPJBYUKXTR' _SCREAMING_SNAKE_CASE = 'FOBHMDKEXQNRAULPGSJVTYICZW' _SCREAMING_SNAKE_CASE = 'ZJXESIUQLHAVRMDOYGTNFWPBKC' # reflector -------------------------- _SCREAMING_SNAKE_CASE = { 'A': 'N', 'N': 'A', 'B': 'O', 'O': 'B', 'C': 'P', 'P': 'C', 'D': 'Q', 'Q': 'D', 'E': 'R', 'R': 'E', 'F': 'S', 'S': 'F', 'G': 'T', 'T': 'G', 'H': 'U', 'U': 'H', 'I': 'V', 'V': 'I', 'J': 'W', 'W': 'J', 'K': 'X', 'X': 'K', 'L': 'Y', 'Y': 'L', 'M': 'Z', 'Z': 'M', } # -------------------------- extra rotors -------------------------- _SCREAMING_SNAKE_CASE = 'RMDJXFUWGISLHVTCQNKYPBEZOA' _SCREAMING_SNAKE_CASE = 'SGLCPQWZHKXAREONTFBVIYJUDM' _SCREAMING_SNAKE_CASE = 'HVSICLTYKQUBXDWAJZOMFGPREN' _SCREAMING_SNAKE_CASE = 'RZWQHFMVDBKICJLNTUXAGYPSOE' _SCREAMING_SNAKE_CASE = 'LFKIJODBEGAMQPXVUHYSTCZRWN' _SCREAMING_SNAKE_CASE = 'KOAEGVDHXPQZMLFTYWJNBRCIUS' def snake_case ( snake_case__ :RotorPositionT , snake_case__ :RotorSelectionT , snake_case__ :str) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(snake_case__))) < 3: _A = F'''Please use 3 unique rotors (not {unique_rotsel})''' raise Exception(snake_case__) # Checks if rotor positions are valid _A , _A , _A = rotpos if not 0 < rotorposa <= len(snake_case__): _A = F'''First rotor position is not within range of 1..26 ({rotorposa}''' raise ValueError(snake_case__) if not 0 < rotorposa <= len(snake_case__): _A = F'''Second rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(snake_case__) if not 0 < rotorposa <= len(snake_case__): _A = F'''Third rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(snake_case__) # Validates string and returns dict _A = _plugboard(snake_case__) return rotpos, rotsel, pbdict def snake_case ( snake_case__ :str) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(snake_case__ , snake_case__): _A = F'''Plugboard setting isn\'t type string ({type(snake_case__)})''' raise TypeError(snake_case__) elif len(snake_case__) % 2 != 0: _A = F'''Odd number of symbols ({len(snake_case__)})''' raise Exception(snake_case__) elif pbstring == "": return {} pbstring.replace(""" """ , """""") # Checks if all characters are unique _A = set() for i in pbstring: if i not in abc: _A = F'''\'{i}\' not in list of symbols''' raise Exception(snake_case__) elif i in tmppbl: _A = F'''Duplicate symbol ({i})''' raise Exception(snake_case__) else: tmppbl.add(snake_case__) del tmppbl # Created the dictionary _A = {} for j in range(0 , len(snake_case__) - 1 , 2): _A = pbstring[j + 1] _A = pbstring[j] return pb def snake_case ( snake_case__ :str , snake_case__ :RotorPositionT , snake_case__ :RotorSelectionT = (rotora, rotora, rotora) , snake_case__ :str = "" , ) -> str: _A = text.upper() _A , _A , _A = _validator( snake_case__ , snake_case__ , plugb.upper()) _A , _A , _A = rotor_position _A , _A , _A = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _A = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _A = plugboard[symbol] # rotor ra -------------------------- _A = abc.index(snake_case__) + rotorposa _A = rotora[index % len(snake_case__)] # rotor rb -------------------------- _A = abc.index(snake_case__) + rotorposa _A = rotora[index % len(snake_case__)] # rotor rc -------------------------- _A = abc.index(snake_case__) + rotorposa _A = rotora[index % len(snake_case__)] # reflector -------------------------- # this is the reason you don't need another machine to decipher _A = reflector[symbol] # 2nd rotors _A = abc[rotora.index(snake_case__) - rotorposa] _A = abc[rotora.index(snake_case__) - rotorposa] _A = abc[rotora.index(snake_case__) - rotorposa] # 2nd plugboard if symbol in plugboard: _A = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(snake_case__): _A = 0 rotorposa += 1 if rotorposa >= len(snake_case__): _A = 0 rotorposa += 1 if rotorposa >= len(snake_case__): _A = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(snake_case__) return "".join(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = 'This is my Python script that emulates the Enigma machine from WWII.' _SCREAMING_SNAKE_CASE = (1, 1, 1) _SCREAMING_SNAKE_CASE = 'pictures' _SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) _SCREAMING_SNAKE_CASE = enigma(message, rotor_pos, rotor_sel, pb) print('Encrypted message:', en) print('Decrypted message:', enigma(en, rotor_pos, rotor_sel, pb))
83
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = '''speech_to_text''' lowerCamelCase :List[str] = ['''past_key_values'''] lowerCamelCase :str = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=12 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=2 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=60_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=2 , lowerCAmelCase_=(5, 5) , lowerCAmelCase_=10_24 , lowerCAmelCase_=80 , lowerCAmelCase_=1 , **lowerCAmelCase_ , ) -> Tuple: _A = vocab_size _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = encoder_layerdrop _A = decoder_layerdrop _A = use_cache _A = encoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = max_source_positions _A = max_target_positions _A = num_conv_layers _A = list(lowerCAmelCase_ ) _A = conv_channels _A = input_feat_per_channel _A = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` """ F'''but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, ''' F'''`config.num_conv_layers = {self.num_conv_layers}`.''' ) super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , )
83
1
import os import time import numpy as np import onnxruntime as ort _SCREAMING_SNAKE_CASE = '1' _SCREAMING_SNAKE_CASE = '0' _SCREAMING_SNAKE_CASE = '1' _SCREAMING_SNAKE_CASE = ort.SessionOptions() _SCREAMING_SNAKE_CASE = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print('Create inference session...') _SCREAMING_SNAKE_CASE = ['TensorrtExecutionProvider', 'CUDAExecutionProvider'] _SCREAMING_SNAKE_CASE = ort.InferenceSession('model.onnx', sess_options=sess_opt, providers=execution_provider) _SCREAMING_SNAKE_CASE = ort.RunOptions() _SCREAMING_SNAKE_CASE = 128 _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) _SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) _SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) print('Warm up phase...') sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('Start inference...') _SCREAMING_SNAKE_CASE = time.time() _SCREAMING_SNAKE_CASE = 2_000 _SCREAMING_SNAKE_CASE = {} for iter in range(max_iters): _SCREAMING_SNAKE_CASE = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('Average Inference Time = {:.3f} ms'.format((time.time() - start_time) * 1_000 / max_iters))
83
from __future__ import annotations from collections.abc import Callable def snake_case ( snake_case__ :Callable[[int | float], int | float] , snake_case__ :int | float , snake_case__ :int | float , snake_case__ :int = 100 , ) -> float: _A = x_start _A = fnc(snake_case__) _A = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area _A = (x_end - x_start) / steps + xa _A = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step _A = xa _A = fxa return area if __name__ == "__main__": def snake_case ( snake_case__ :Tuple) -> List[str]: return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') _SCREAMING_SNAKE_CASE = 10 while i <= 100_000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
83
1
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class a ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = VideoToVideoSDPipeline lowerCamelCase :Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'''video'''} ) - {'''image''', '''width''', '''height'''} lowerCamelCase :Dict = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''video'''} ) - {'''image'''} lowerCamelCase :Union[str, Any] = PipelineTesterMixin.required_optional_params - {'''latents'''} lowerCamelCase :int = False # No `output_type`. lowerCamelCase :Union[str, Any] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def UpperCAmelCase ( self ) -> Optional[Any]: torch.manual_seed(0 ) _A = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _A = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=lowerCAmelCase_ , set_alpha_to_one=lowerCAmelCase_ , ) torch.manual_seed(0 ) _A = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) _A = CLIPTextModel(lowerCAmelCase_ ) _A = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _A = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=0 ) -> Optional[Any]: # 3 frames _A = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) if str(lowerCAmelCase_ ).startswith("""mps""" ): _A = torch.manual_seed(lowerCAmelCase_ ) else: _A = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _A = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCAmelCase ( self ) -> Dict: _A = """cpu""" # ensure determinism for the device-dependent torch.Generator _A = self.get_dummy_components() _A = VideoToVideoSDPipeline(**lowerCAmelCase_ ) _A = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A = self.get_dummy_inputs(lowerCAmelCase_ ) _A = """np""" _A = sd_pipe(**lowerCAmelCase_ ).frames _A = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _A = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCAmelCase ( self ) -> List[str]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase_ , expected_max_diff=5E-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCAmelCase ( self ) -> Tuple: pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCAmelCase ( self ) -> Optional[Any]: pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCAmelCase ( self ) -> int: pass def UpperCAmelCase ( self ) -> int: return super().test_progress_bar() @slow @skip_mps class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: _A = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames _A = torch.Generator(device="""cpu""" ).manual_seed(0 ) _A = torch.randn((1, 10, 3, 10_24, 5_76) , generator=lowerCAmelCase_ ) _A = video.to("""cuda""" ) _A = """Spiderman is surfing""" _A = pipe(lowerCAmelCase_ , video=lowerCAmelCase_ , generator=lowerCAmelCase_ , num_inference_steps=3 , output_type="""pt""" ).frames _A = np.array([-1.045_8984, -1.127_9297, -0.966_3086, -0.9150_3906, -0.7509_7656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1E-2
83
import numpy as np import qiskit def snake_case ( snake_case__ :int = 8 , snake_case__ :int | None = None) -> str: _A = np.random.default_rng(seed=snake_case__) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _A = 6 * key_len # Measurement basis for Alice's qubits. _A = rng.integers(2 , size=snake_case__) # The set of states Alice will prepare. _A = rng.integers(2 , size=snake_case__) # Measurement basis for Bob's qubits. _A = rng.integers(2 , size=snake_case__) # Quantum Circuit to simulate BB84 _A = qiskit.QuantumCircuit(snake_case__ , name="""BB84""") # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case__): if alice_state[index] == 1: bbaa_circ.x(snake_case__) if alice_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case__): if bob_basis[index] == 1: bbaa_circ.h(snake_case__) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _A = qiskit.Aer.get_backend("""aer_simulator""") # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _A = qiskit.execute(snake_case__ , snake_case__ , shots=1 , seed_simulator=snake_case__) # Returns the result of measurement. _A = job.result().get_counts(snake_case__).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _A = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case__ , snake_case__ , snake_case__) if alice_basis_bit == bob_basis_bit ]) # Get final key. Pad with 0 if too short, otherwise truncate. _A = gen_key[:key_len] if len(snake_case__) >= key_len else gen_key.ljust(snake_case__ , """0""") return key if __name__ == "__main__": print(F'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
83
1
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'microsoft/conditional-detr-resnet-50': ( 'https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json' ), } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Dict = '''conditional_detr''' lowerCamelCase :Tuple = ['''past_key_values'''] lowerCamelCase :List[Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , lowerCAmelCase_=True , lowerCAmelCase_=None , lowerCAmelCase_=3 , lowerCAmelCase_=3_00 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=8 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=8 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=1.0 , lowerCAmelCase_=False , lowerCAmelCase_="sine" , lowerCAmelCase_="resnet50" , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=2 , lowerCAmelCase_=5 , lowerCAmelCase_=2 , lowerCAmelCase_=1 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , lowerCAmelCase_=5 , lowerCAmelCase_=2 , lowerCAmelCase_=0.25 , **lowerCAmelCase_ , ) -> Optional[Any]: if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) _A = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = backbone_config.get("""model_type""" ) _A = CONFIG_MAPPING[backbone_model_type] _A = config_class.from_dict(lowerCAmelCase_ ) _A = use_timm_backbone _A = backbone_config _A = num_channels _A = num_queries _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = init_xavier_std _A = encoder_layerdrop _A = decoder_layerdrop _A = encoder_layers _A = auxiliary_loss _A = position_embedding_type _A = backbone _A = use_pretrained_backbone _A = dilation # Hungarian matcher _A = class_cost _A = bbox_cost _A = giou_cost # Loss coefficients _A = mask_loss_coefficient _A = dice_loss_coefficient _A = cls_loss_coefficient _A = bbox_loss_coefficient _A = giou_loss_coefficient _A = focal_alpha super().__init__(is_encoder_decoder=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def UpperCAmelCase ( self ) -> int: return self.encoder_attention_heads @property def UpperCAmelCase ( self ) -> int: return self.d_model def UpperCAmelCase ( self ) -> List[Any]: _A = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _A = self.backbone_config.to_dict() _A = self.__class__.model_type return output class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = version.parse('''1.11''' ) @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCAmelCase ( self ) -> float: return 1E-5 @property def UpperCAmelCase ( self ) -> int: return 12
83
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def snake_case ( snake_case__ :int) -> Optional[int]: return EnvironmentCommand() def snake_case ( snake_case__ :Tuple) -> List[str]: return EnvironmentCommand(args.accelerate_config_file) class a ( __lowerCAmelCase ): """simple docstring""" @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: _A = parser.add_parser("""env""" ) download_parser.set_defaults(func=lowerCAmelCase_ ) download_parser.add_argument( """--accelerate-config_file""" , default=lowerCAmelCase_ , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=lowerCAmelCase_ ) def __init__( self , lowerCAmelCase_ , *lowerCAmelCase_ ) -> None: _A = accelerate_config_file def UpperCAmelCase ( self ) -> Dict: _A = """not installed""" if is_safetensors_available(): import safetensors _A = safetensors.__version__ elif importlib.util.find_spec("""safetensors""" ) is not None: import safetensors _A = F'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' _A = """not installed""" _A = _A = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file _A = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(lowerCAmelCase_ ): _A = load_config_from_file(self._accelerate_config_file ).to_dict() _A = ( """\n""".join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else F'''\t{accelerate_config}''' ) _A = """not installed""" _A = """NA""" if is_torch_available(): import torch _A = torch.__version__ _A = torch.cuda.is_available() _A = """not installed""" _A = """NA""" if is_tf_available(): import tensorflow as tf _A = tf.__version__ try: # deprecated in v2.1 _A = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool _A = bool(tf.config.list_physical_devices("""GPU""" ) ) _A = """not installed""" _A = """not installed""" _A = """not installed""" _A = """NA""" if is_flax_available(): import flax import jax import jaxlib _A = flax.__version__ _A = jax.__version__ _A = jaxlib.__version__ _A = jax.lib.xla_bridge.get_backend().platform _A = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": F'''{safetensors_version}''', """Accelerate version""": F'''{accelerate_version}''', """Accelerate config""": F'''{accelerate_config_str}''', """PyTorch version (GPU?)""": F'''{pt_version} ({pt_cuda_available})''', """Tensorflow version (GPU?)""": F'''{tf_version} ({tf_cuda_available})''', """Flax version (CPU?/GPU?/TPU?)""": F'''{flax_version} ({jax_backend})''', """Jax version""": F'''{jax_version}''', """JaxLib version""": F'''{jaxlib_version}''', """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""" ) print(self.format_dict(lowerCAmelCase_ ) ) return info @staticmethod def UpperCAmelCase ( lowerCAmelCase_ ) -> Tuple: return "\n".join([F'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
83
1
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case ( snake_case__ :str) -> int: _A = torch.load(snake_case__ , map_location="""cpu""") if "model" in sd.keys(): _A = torch.load(snake_case__ , map_location="""cpu""")["""model"""] # pop unnecessary weights _A = [ """decoder.version""", """decoder.output_projection.weight""", ] for key in keys_to_delete: if key in sd: sd.pop(snake_case__) _A = { """decoder.project_in_dim.weight""": """decoder.project_in.weight""", """decoder.project_out_dim.weight""": """decoder.project_out.weight""", """decoder.layer_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.layer_norm.bias""": """decoder.final_layer_norm.bias""", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: _A = sd.pop(snake_case__) _A = list(sd.keys()) for key in keys: if ".qkv_proj." in key: _A = sd[key] # We split QKV in separate Q,K,V _A = key.replace(""".qkv_proj.""" , """.q_proj.""") _A = key.replace(""".qkv_proj.""" , """.k_proj.""") _A = key.replace(""".qkv_proj.""" , """.v_proj.""") _A = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 _A , _A , _A = torch.split(snake_case__ , depth // 3 , dim=0) _A = q _A = k _A = v del sd[key] return sd @torch.no_grad() def snake_case ( snake_case__ :List[Any] , snake_case__ :Tuple , snake_case__ :List[str]=None) -> Tuple: _A = load_checkpoint(snake_case__) if config is not None: _A = OPTConfig.from_pretrained(snake_case__) else: _A = OPTConfig() _A = OPTModel(snake_case__).half().eval() model.load_state_dict(snake_case__) # Check results Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '--fairseq_path', type=str, help=( 'path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:' ' https://huggingface.co/models?other=opt_metasq' ), ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--hf_config', default=None, type=str, help='Define HF config.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
83
import colorsys from PIL import Image # type: ignore def snake_case ( snake_case__ :float , snake_case__ :float , snake_case__ :int) -> float: _A = x _A = y for step in range(snake_case__): # noqa: B007 _A = a * a - b * b + x _A = 2 * a * b + y _A = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def snake_case ( snake_case__ :float) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255) for i in colorsys.hsv_to_rgb(snake_case__ , 1 , 1)) def snake_case ( snake_case__ :int = 800 , snake_case__ :int = 600 , snake_case__ :float = -0.6 , snake_case__ :float = 0 , snake_case__ :float = 3.2 , snake_case__ :int = 50 , snake_case__ :bool = True , ) -> Image.Image: _A = Image.new("""RGB""" , (image_width, image_height)) _A = img.load() # loop through the image-coordinates for image_x in range(snake_case__): for image_y in range(snake_case__): # determine the figure-coordinates based on the image-coordinates _A = figure_width / image_width * image_height _A = figure_center_x + (image_x / image_width - 0.5) * figure_width _A = figure_center_y + (image_y / image_height - 0.5) * figure_height _A = get_distance(snake_case__ , snake_case__ , snake_case__) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _A = get_color_coded_rgb(snake_case__) else: _A = get_black_and_white_rgb(snake_case__) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _SCREAMING_SNAKE_CASE = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
1
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = R'\n Args:\n input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax\n or scores for each vocabulary token after SoftMax.\n kwargs (`Dict[str, Any]`, *optional*):\n Additional stopping criteria specific kwargs.\n\n Return:\n `bool`. `False` indicates we should continue, `True` indicates we should stop.\n\n' class a ( __lowerCAmelCase ): """simple docstring""" @add_start_docstrings(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> bool: raise NotImplementedError("""StoppingCriteria needs to be subclassed""" ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> List[str]: _A = max_length _A = max_position_embeddings @add_start_docstrings(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> bool: _A = input_ids.shape[-1] _A = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( """This is a friendly reminder - the current text generation call will exceed the model's predefined """ F'''maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ''' """exceptions, performance degradation, or nothing at all.""" ) return is_done class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ ) -> Dict: warnings.warn( """The class `MaxNewTokensCriteria` is deprecated. """ F'''Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ''' """with `max_length = start_length + max_new_tokens` instead.""" , lowerCAmelCase_ , ) _A = start_length _A = max_new_tokens _A = start_length + max_new_tokens @add_start_docstrings(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> bool: return input_ids.shape[-1] >= self.max_length class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Any: _A = max_time _A = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> bool: return time.time() - self.initial_timestamp > self.max_time class a ( __lowerCAmelCase ): """simple docstring""" @add_start_docstrings(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) -> bool: return any(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) for criteria in self ) @property def UpperCAmelCase ( self ) -> Optional[int]: for stopping_criterium in self: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return stopping_criterium.max_length elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return stopping_criterium.max_length return None def snake_case ( snake_case__ :StoppingCriteriaList , snake_case__ :int) -> StoppingCriteriaList: _A = stopping_criteria.max_length _A = deepcopy(snake_case__) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn("""You set different `max_length` for stopping criteria and `max_length` parameter""" , snake_case__) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=snake_case__)) return new_stopping_criteria
83
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _SCREAMING_SNAKE_CASE = datasets.logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _SCREAMING_SNAKE_CASE = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _SCREAMING_SNAKE_CASE = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def snake_case ( snake_case__ :Optional[Any] , snake_case__ :str , snake_case__ :List[str]=False , snake_case__ :Dict=False , snake_case__ :Any=True , snake_case__ :List[str]=False , snake_case__ :Optional[Any]="dummy_doc") -> List[Any]: _A = {doc: key_lines} _A = {doc: sys_lines} _A = {} _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A = 0 _A , _A = reader.get_doc_mentions(snake_case__ , key_doc_lines[doc] , snake_case__) key_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) _A , _A = reader.get_doc_mentions(snake_case__ , sys_doc_lines[doc] , snake_case__) sys_singletons_num += singletons_num if NP_only or min_span: _A = reader.set_annotated_parse_trees(snake_case__ , key_doc_lines[doc] , snake_case__ , snake_case__) if remove_nested: _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _A , _A = reader.remove_nested_coref_mentions(snake_case__ , snake_case__) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = reader.get_mention_assignments(snake_case__ , snake_case__) _A = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F'''annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}''') logger.info( """Number of resulting singleton clusters in the key """ F'''annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}''') if not keep_singletons: logger.info( F'''{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ''' """files, respectively""") return doc_coref_infos def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Dict , snake_case__ :Tuple , snake_case__ :Dict , snake_case__ :Tuple) -> int: _A = get_coref_infos(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) _A = {} _A = 0 _A = 0 for name, metric in metrics: _A , _A , _A = evaluator.evaluate_documents(snake_case__ , snake_case__ , beta=1) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'''{name}/recall''': recall, F'''{name}/precision''': precision, F'''{name}/f1''': fa}) logger.info( name.ljust(10) , F'''Recall: {recall * 100:.2f}''' , F''' Precision: {precision * 100:.2f}''' , F''' F1: {fa * 100:.2f}''' , ) if conll_subparts_num == 3: _A = (conll / 3) * 100 logger.info(F'''CoNLL score: {conll:.2f}''') output_scores.update({"""conll_score""": conll}) return output_scores def snake_case ( snake_case__ :Union[str, Any]) -> List[Any]: _A = False for line in key_lines: if not line.startswith("""#"""): if len(line.split()) > 6: _A = line.split()[5] if not parse_col == "-": _A = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Union[str, Any]: _A = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _A = util.check_gold_parse_annotation(lowerCAmelCase_ ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _A = evaluate( key_lines=lowerCAmelCase_ , sys_lines=lowerCAmelCase_ , metrics=lowerCAmelCase_ , NP_only=lowerCAmelCase_ , remove_nested=lowerCAmelCase_ , keep_singletons=lowerCAmelCase_ , min_span=lowerCAmelCase_ , ) return score
83
1
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _SCREAMING_SNAKE_CASE = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. _SCREAMING_SNAKE_CASE = direct_transformers_import(PATH_TO_TRANSFORMERS) _SCREAMING_SNAKE_CASE = transformers.models.auto.configuration_auto.CONFIG_MAPPING _SCREAMING_SNAKE_CASE = { # used to compute the property `self.chunk_length` 'EncodecConfig': ['overlap'], # used as `self.bert_model = BertModel(config, ...)` 'DPRConfig': True, # not used in modeling files, but it's an important information 'FSMTConfig': ['langs'], # used internally in the configuration class file 'GPTNeoConfig': ['attention_types'], # used internally in the configuration class file 'EsmConfig': ['is_folding_model'], # used during training (despite we don't have training script for these models yet) 'Mask2FormerConfig': ['ignore_value'], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) 'OneFormerConfig': ['ignore_value', 'norm'], # used during preprocessing and collation, see `collating_graphormer.py` 'GraphormerConfig': ['spatial_pos_max'], # used internally in the configuration class file 'T5Config': ['feed_forward_proj'], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally 'MT5Config': ['feed_forward_proj', 'tokenizer_class'], 'UMT5Config': ['feed_forward_proj', 'tokenizer_class'], # used internally in the configuration class file 'LongT5Config': ['feed_forward_proj'], # used internally in the configuration class file 'SwitchTransformersConfig': ['feed_forward_proj'], # having default values other than `1e-5` - we can't fix them without breaking 'BioGptConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'GLPNConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'SegformerConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'CvtConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'PerceiverConfig': ['layer_norm_eps'], # used internally to calculate the feature size 'InformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'TimeSeriesTransformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'AutoformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate `mlp_dim` 'SamVisionConfig': ['mlp_ratio'], # For (head) training, but so far not implemented 'ClapAudioConfig': ['num_classes'], # Not used, but providing useful information to users 'SpeechT5HifiGanConfig': ['sampling_rate'], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { 'CLIPSegConfig': True, 'DeformableDetrConfig': True, 'DetaConfig': True, 'DinatConfig': True, 'DonutSwinConfig': True, 'EfficientFormerConfig': True, 'FSMTConfig': True, 'JukeboxConfig': True, 'LayoutLMv2Config': True, 'MaskFormerSwinConfig': True, 'MT5Config': True, 'NatConfig': True, 'OneFormerConfig': True, 'PerceiverConfig': True, 'RagConfig': True, 'SpeechT5Config': True, 'SwinConfig': True, 'Swin2SRConfig': True, 'Swinv2Config': True, 'SwitchTransformersConfig': True, 'TableTransformerConfig': True, 'TapasConfig': True, 'TransfoXLConfig': True, 'UniSpeechConfig': True, 'UniSpeechSatConfig': True, 'WavLMConfig': True, 'WhisperConfig': True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) 'JukeboxPriorConfig': True, # TODO: @Younes (for `is_decoder`) 'Pix2StructTextConfig': True, } ) def snake_case ( snake_case__ :Dict , snake_case__ :Tuple , snake_case__ :Any , snake_case__ :Any) -> List[str]: _A = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( F'''config.{attribute}''' in modeling_source or F'''getattr(config, "{attribute}"''' in modeling_source or F'''getattr(self.config, "{attribute}"''' in modeling_source ): _A = True # Deal with multi-line cases elif ( re.search( RF'''getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*"{attribute}"''' , snake_case__ , ) is not None ): _A = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: _A = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files _A = [ """bos_index""", """eos_index""", """pad_index""", """unk_index""", """mask_index""", """image_size""", """use_cache""", """out_features""", """out_indices""", ] _A = ["""encoder_no_repeat_ngram_size"""] # Special cases to be allowed _A = True if not attribute_used: _A = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: _A = True elif attribute in ["tie_word_embeddings"] and default_value is False: _A = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: _A = True elif attribute.endswith("""_token_id"""): _A = True # configuration class specific cases if not case_allowed: _A = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , []) _A = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def snake_case ( snake_case__ :Dict) -> List[str]: _A = dict(inspect.signature(config_class.__init__).parameters) _A = [x for x in list(signature.keys()) if x not in ["""self""", """kwargs"""]] _A = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass _A = {} if len(config_class.attribute_map) > 0: _A = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files _A = inspect.getsourcefile(snake_case__) _A = os.path.dirname(snake_case__) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. _A = [os.path.join(snake_case__ , snake_case__) for fn in os.listdir(snake_case__) if fn.startswith("""modeling_""")] # Get the source code strings _A = [] for path in modeling_paths: if os.path.isfile(snake_case__): with open(snake_case__) as fp: modeling_sources.append(fp.read()) _A = [] for config_param, default_value in zip(snake_case__ , snake_case__): # `attributes` here is all the variant names for `config_param` _A = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param]) if not check_attribute_being_used(snake_case__ , snake_case__ , snake_case__ , snake_case__): unused_attributes.append(attributes[0]) return sorted(snake_case__) def snake_case ( ) -> Union[str, Any]: _A = {} for _config_class in list(CONFIG_MAPPING.values()): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) _A = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class) , lambda snake_case__: inspect.isclass(snake_case__) and issubclass(snake_case__ , snake_case__) and inspect.getmodule(snake_case__) == inspect.getmodule(_config_class) , ) ] for config_class in config_classes_in_module: _A = check_config_attributes_being_used(snake_case__) if len(snake_case__) > 0: _A = unused_attributes if len(snake_case__) > 0: _A = """The following configuration classes contain unused attributes in the corresponding modeling files:\n""" for name, attributes in configs_with_unused_attributes.items(): error += F'''{name}: {attributes}\n''' raise ValueError(snake_case__) if __name__ == "__main__": check_config_attributes()
83
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } _SCREAMING_SNAKE_CASE = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } _SCREAMING_SNAKE_CASE = {'facebook/blenderbot_small-90M': 512} def snake_case ( snake_case__ :Tuple) -> str: _A = set() _A = word[0] for char in word[1:]: pairs.add((prev_char, char)) _A = char _A = set(snake_case__) return pairs class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :List[Any] = VOCAB_FILES_NAMES lowerCamelCase :Tuple = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase :List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase :int = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="__start__" , lowerCAmelCase_="__end__" , lowerCAmelCase_="__unk__" , lowerCAmelCase_="__null__" , **lowerCAmelCase_ , ) -> int: super().__init__(unk_token=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , **lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _A = json.load(lowerCAmelCase_ ) _A = {v: k for k, v in self.encoder.items()} with open(lowerCAmelCase_ , encoding="""utf-8""" ) as merges_handle: _A = merges_handle.read().split("""\n""" )[1:-1] _A = [tuple(merge.split() ) for merge in merges] _A = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_ ) ) ) ) _A = {} @property def UpperCAmelCase ( self ) -> int: return len(self.encoder ) def UpperCAmelCase ( self ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: if token in self.cache: return self.cache[token] _A = re.sub("""([.,!?()])""" , r""" \1""" , lowerCAmelCase_ ) _A = re.sub("""(')""" , r""" \1 """ , lowerCAmelCase_ ) _A = re.sub(r"""\s{2,}""" , """ """ , lowerCAmelCase_ ) if "\n" in token: _A = token.replace("""\n""" , """ __newln__""" ) _A = token.split(""" """ ) _A = [] for token in tokens: if not len(lowerCAmelCase_ ): continue _A = token.lower() _A = tuple(lowerCAmelCase_ ) _A = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) _A = get_pairs(lowerCAmelCase_ ) if not pairs: words.append(lowerCAmelCase_ ) continue while True: _A = min(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : self.bpe_ranks.get(lowerCAmelCase_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break _A , _A = bigram _A = [] _A = 0 while i < len(lowerCAmelCase_ ): try: _A = word.index(lowerCAmelCase_ , lowerCAmelCase_ ) new_word.extend(word[i:j] ) _A = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A = tuple(lowerCAmelCase_ ) _A = new_word if len(lowerCAmelCase_ ) == 1: break else: _A = get_pairs(lowerCAmelCase_ ) _A = """@@ """.join(lowerCAmelCase_ ) _A = word[:-4] _A = word words.append(lowerCAmelCase_ ) return " ".join(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[str]: _A = [] _A = re.findall(r"""\S+\n?""" , lowerCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(lowerCAmelCase_ ).split(""" """ ) ) ) return split_tokens def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: _A = token.lower() return self.encoder.get(lowerCAmelCase_ , self.encoder.get(self.unk_token ) ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: return self.decoder.get(lowerCAmelCase_ , self.unk_token ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = """ """.join(lowerCAmelCase_ ).replace("""@@ """ , """""" ).strip() return out_string def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) _A = os.path.join( lowerCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase_ , ensure_ascii=lowerCAmelCase_ ) + """\n""" ) _A = 0 with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase_ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) _A = token_index writer.write(""" """.join(lowerCAmelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file
83
1
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def snake_case ( snake_case__ :Union[str, Any]="") -> str: _A = tempfile.mkdtemp() return os.path.join(snake_case__ , str(uuid.uuida()) + suffix) @require_soundfile @require_torch class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> int: _A = torch.rand(12 , dtype=torch.floataa ) - 0.5 _A = AgentAudio(lowerCAmelCase_ ) _A = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase_ , agent_type.to_raw() , atol=1E-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowerCAmelCase_ ) ) # Ensure that the file contains the same value as the original tensor _A , _A = sf.read(lowerCAmelCase_ ) self.assertTrue(torch.allclose(lowerCAmelCase_ , torch.tensor(lowerCAmelCase_ ) , atol=1E-4 ) ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = torch.rand(12 , dtype=torch.floataa ) - 0.5 _A = get_new_path(suffix=""".wav""" ) sf.write(lowerCAmelCase_ , lowerCAmelCase_ , 1_60_00 ) _A = AgentAudio(lowerCAmelCase_ ) self.assertTrue(torch.allclose(lowerCAmelCase_ , agent_type.to_raw() , atol=1E-4 ) ) self.assertEqual(agent_type.to_string() , lowerCAmelCase_ ) @require_vision @require_torch class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Any: _A = torch.randint(0 , 2_56 , (64, 64, 3) ) _A = AgentImage(lowerCAmelCase_ ) _A = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase_ , agent_type._tensor , atol=1E-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase_ ) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = Path(get_tests_dir("""fixtures/tests_samples/COCO""" ) ) / """000000039769.png""" _A = Image.open(lowerCAmelCase_ ) _A = AgentImage(lowerCAmelCase_ ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase_ ) ) def UpperCAmelCase ( self ) -> List[str]: _A = Path(get_tests_dir("""fixtures/tests_samples/COCO""" ) ) / """000000039769.png""" _A = Image.open(lowerCAmelCase_ ) _A = AgentImage(lowerCAmelCase_ ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase_ ) ) class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Optional[Any]: _A = """Hey!""" _A = AgentText(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , agent_type.to_string() ) self.assertEqual(lowerCAmelCase_ , agent_type.to_raw() ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ )
83
_SCREAMING_SNAKE_CASE = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on _SCREAMING_SNAKE_CASE = {value: key for key, value in MORSE_CODE_DICT.items()} def snake_case ( snake_case__ :str) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper()) def snake_case ( snake_case__ :str) -> str: return "".join(REVERSE_DICT[char] for char in message.split()) def snake_case ( ) -> None: _A = """Morse code here!""" print(snake_case__) _A = encrypt(snake_case__) print(snake_case__) _A = decrypt(snake_case__) print(snake_case__) if __name__ == "__main__": main()
83
1
import numpy as np _SCREAMING_SNAKE_CASE = [ ['a', 'b', 'c', 'd', 'e'], ['f', 'g', 'h', 'i', 'k'], ['l', 'm', 'n', 'o', 'p'], ['q', 'r', 's', 't', 'u'], ['v', 'w', 'x', 'y', 'z'], ] class a : """simple docstring""" def __init__( self ) -> None: _A = np.array(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> np.ndarray: _A , _A = np.where(letter == self.SQUARE ) _A = np.concatenate([indexa + 1, indexa + 1] ) return indexes def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = self.SQUARE[indexa - 1, indexa - 1] return letter def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = message.lower() _A = message.replace(""" """ , """""" ) _A = message.replace("""j""" , """i""" ) _A = np.empty((2, len(lowerCAmelCase_ )) ) for letter_index in range(len(lowerCAmelCase_ ) ): _A = self.letter_to_numbers(message[letter_index] ) _A = numbers[0] _A = numbers[1] _A = first_step.reshape(2 * len(lowerCAmelCase_ ) ) _A = """""" for numbers_index in range(len(lowerCAmelCase_ ) ): _A = int(second_step[numbers_index * 2] ) _A = int(second_step[(numbers_index * 2) + 1] ) _A = self.numbers_to_letter(lowerCAmelCase_ , lowerCAmelCase_ ) _A = encoded_message + letter return encoded_message def UpperCAmelCase ( self , lowerCAmelCase_ ) -> str: _A = message.lower() message.replace(""" """ , """""" ) _A = np.empty(2 * len(lowerCAmelCase_ ) ) for letter_index in range(len(lowerCAmelCase_ ) ): _A = self.letter_to_numbers(message[letter_index] ) _A = numbers[0] _A = numbers[1] _A = first_step.reshape((2, len(lowerCAmelCase_ )) ) _A = """""" for numbers_index in range(len(lowerCAmelCase_ ) ): _A = int(second_step[0, numbers_index] ) _A = int(second_step[1, numbers_index] ) _A = self.numbers_to_letter(lowerCAmelCase_ , lowerCAmelCase_ ) _A = decoded_message + letter return decoded_message
83
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _SCREAMING_SNAKE_CASE = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('0.12.2'): raise Exception('requires fairseq >= 0.12.2') if version.parse(fairseq.__version__) > version.parse('2'): raise Exception('requires fairseq < v2') logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = 'Hello, World!' _SCREAMING_SNAKE_CASE = 'en_XX' def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :bool) -> Tuple: _A = Path("""data_bin""") _A = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(snake_case__).parent) , checkpoint_file=Path(snake_case__).name , _name="""xmod_base""" , arch="""xmod_base""" , task="""multilingual_masked_lm""" , data_name_or_path=str(snake_case__) , bpe="""sentencepiece""" , sentencepiece_model=str(Path(snake_case__).parent / """sentencepiece.bpe.model""") , src_dict=str(data_dir / """dict.txt""") , ) xmod.eval() # disable dropout print(snake_case__) _A = xmod.model.encoder.sentence_encoder _A = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1E-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , """bottleneck""" , 2) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: _A = xmod.model.classification_heads["""mnli"""].out_proj.weight.shape[0] print("""Our X-MOD config:""" , snake_case__) _A = XmodForSequenceClassification(snake_case__) if classification_head else XmodForMaskedLM(snake_case__) model.eval() # Now let's copy all the weights. # Embeddings _A = xmod_sent_encoder.embed_tokens.weight _A = xmod_sent_encoder.embed_positions.weight _A = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight) # just zero them out b/c xmod doesn't use them. _A = xmod_sent_encoder.layernorm_embedding.weight _A = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers): # Encoder: start of layer _A = model.roberta.encoder.layer[i] _A = xmod_sent_encoder.layers[i] # self attention _A = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size)) ): raise AssertionError("""Dimensions of self-attention weights do not match.""") _A = xmod_layer.self_attn.q_proj.weight _A = xmod_layer.self_attn.q_proj.bias _A = xmod_layer.self_attn.k_proj.weight _A = xmod_layer.self_attn.k_proj.bias _A = xmod_layer.self_attn.v_proj.weight _A = xmod_layer.self_attn.v_proj.bias # self-attention output _A = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError("""Dimensions of self-attention output weights do not match.""") _A = xmod_layer.self_attn.out_proj.weight _A = xmod_layer.self_attn.out_proj.bias _A = xmod_layer.self_attn_layer_norm.weight _A = xmod_layer.self_attn_layer_norm.bias # intermediate _A = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("""Dimensions of intermediate weights do not match.""") _A = xmod_layer.fca.weight _A = xmod_layer.fca.bias # output _A = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("""Dimensions of feed-forward weights do not match.""") _A = xmod_layer.fca.weight _A = xmod_layer.fca.bias _A = xmod_layer.final_layer_norm.weight _A = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: _A = xmod_layer.adapter_layer_norm.weight _A = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys()) != sorted(xmod_layer.adapter_modules.keys()): raise AssertionError("""Lists of language adapters do not match.""") for lang_code, adapter in xmod_layer.adapter_modules.items(): _A = bert_output.adapter_modules[lang_code] _A = xmod_layer.adapter_modules[lang_code] _A = from_adapter.fca.weight _A = from_adapter.fca.bias _A = from_adapter.fca.weight _A = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: _A = xmod_sent_encoder.layer_norm.weight _A = xmod_sent_encoder.layer_norm.bias if classification_head: _A = xmod.model.classification_heads["""mnli"""].dense.weight _A = xmod.model.classification_heads["""mnli"""].dense.bias _A = xmod.model.classification_heads["""mnli"""].out_proj.weight _A = xmod.model.classification_heads["""mnli"""].out_proj.bias else: # LM Head _A = xmod.model.encoder.lm_head.dense.weight _A = xmod.model.encoder.lm_head.dense.bias _A = xmod.model.encoder.lm_head.layer_norm.weight _A = xmod.model.encoder.lm_head.layer_norm.bias _A = xmod.model.encoder.lm_head.weight _A = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. _A = xmod.encode(snake_case__).unsqueeze(0) # batch of size 1 model.roberta.set_default_language(snake_case__) _A = model(snake_case__)[0] if classification_head: _A = xmod.model.classification_heads["""mnli"""](xmod.extract_features(snake_case__)) else: _A = xmod.model(snake_case__ , lang_id=[SAMPLE_LANGUAGE])[0] print(our_output.shape , their_output.shape) _A = torch.max(torch.abs(our_output - their_output)).item() print(F'''max_absolute_diff = {max_absolute_diff}''') # ~ 1e-7 _A = torch.allclose(snake_case__ , snake_case__ , atol=1E-3) print("""Do both models output the same tensors?""" , """🔥""" if success else """💩""") if not success: raise Exception("""Something went wRoNg""") Path(snake_case__).mkdir(parents=snake_case__ , exist_ok=snake_case__) print(F'''Saving model to {pytorch_dump_folder_path}''') model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xmod_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
83
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Tuple = '''philschmid/bart-large-cnn-samsum''' lowerCamelCase :Tuple = ( '''This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ''' '''and returns a summary of the text.''' ) lowerCamelCase :List[Any] = '''summarizer''' lowerCamelCase :List[str] = AutoTokenizer lowerCamelCase :Dict = AutoModelForSeqaSeqLM lowerCamelCase :int = ['''text'''] lowerCamelCase :List[Any] = ['''text'''] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> List[Any]: return self.pre_processor(lowerCAmelCase_ , return_tensors="""pt""" , truncation=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: return self.model.generate(**lowerCAmelCase_ )[0] def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: return self.pre_processor.decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ )
83
1
import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class a ( tf.keras.optimizers.schedules.LearningRateSchedule ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1.0 , lowerCAmelCase_ = None , ) -> Union[str, Any]: super().__init__() _A = initial_learning_rate _A = warmup_steps _A = power _A = decay_schedule_fn _A = name def __call__( self , lowerCAmelCase_ ) -> Optional[int]: with tf.name_scope(self.name or """WarmUp""" ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. _A = tf.cast(lowerCAmelCase_ , tf.floataa ) _A = tf.cast(self.warmup_steps , tf.floataa ) _A = global_step_float / warmup_steps_float _A = self.initial_learning_rate * tf.math.pow(lowerCAmelCase_ , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=lowerCAmelCase_ , ) def UpperCAmelCase ( self ) -> List[Any]: return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def snake_case ( snake_case__ :float , snake_case__ :int , snake_case__ :int , snake_case__ :float = 0.0 , snake_case__ :float = 0.9 , snake_case__ :float = 0.999 , snake_case__ :float = 1E-8 , snake_case__ :Optional[float] = None , snake_case__ :Optional[float] = None , snake_case__ :float = 0.0 , snake_case__ :float = 1.0 , snake_case__ :Optional[List[str]] = None , ) -> List[Any]: _A = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=snake_case__ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=snake_case__ , ) if num_warmup_steps: _A = WarmUp( initial_learning_rate=snake_case__ , decay_schedule_fn=snake_case__ , warmup_steps=snake_case__ , ) if weight_decay_rate > 0.0: _A = AdamWeightDecay( learning_rate=snake_case__ , weight_decay_rate=snake_case__ , beta_a=snake_case__ , beta_a=snake_case__ , epsilon=snake_case__ , clipnorm=snake_case__ , global_clipnorm=snake_case__ , exclude_from_weight_decay=["""LayerNorm""", """layer_norm""", """bias"""] , include_in_weight_decay=snake_case__ , ) else: _A = tf.keras.optimizers.Adam( learning_rate=snake_case__ , beta_a=snake_case__ , beta_a=snake_case__ , epsilon=snake_case__ , clipnorm=snake_case__ , global_clipnorm=snake_case__ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ = 0.001 , lowerCAmelCase_ = 0.9 , lowerCAmelCase_ = 0.999 , lowerCAmelCase_ = 1E-7 , lowerCAmelCase_ = False , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = "AdamWeightDecay" , **lowerCAmelCase_ , ) -> List[str]: super().__init__(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) _A = weight_decay_rate _A = include_in_weight_decay _A = exclude_from_weight_decay @classmethod def UpperCAmelCase ( cls , lowerCAmelCase_ ) -> str: _A = {"""WarmUp""": WarmUp} return super(lowerCAmelCase_ , cls ).from_config(lowerCAmelCase_ , custom_objects=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[Any]: super(lowerCAmelCase_ , self )._prepare_local(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _A = tf.constant( self.weight_decay_rate , name="""adam_weight_decay_rate""" ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]["""weight_decay_rate"""] , use_locking=self._use_locking , ) return tf.no_op() def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None , **lowerCAmelCase_ ) -> Any: _A , _A = list(zip(*lowerCAmelCase_ ) ) return super(lowerCAmelCase_ , self ).apply_gradients(zip(lowerCAmelCase_ , lowerCAmelCase_ ) , name=lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: if apply_state is None: return self._decayed_lr_t[var_dtype], {} _A = apply_state or {} _A = apply_state.get((var_device, var_dtype) ) if coefficients is None: _A = self._fallback_apply_state(lowerCAmelCase_ , lowerCAmelCase_ ) _A = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ) -> str: _A , _A = self._get_lr(var.device , var.dtype.base_dtype , lowerCAmelCase_ ) _A = self._decay_weights_op(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) with tf.control_dependencies([decay] ): return super(lowerCAmelCase_ , self )._resource_apply_dense(lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ) -> Optional[Any]: _A , _A = self._get_lr(var.device , var.dtype.base_dtype , lowerCAmelCase_ ) _A = self._decay_weights_op(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) with tf.control_dependencies([decay] ): return super(lowerCAmelCase_ , self )._resource_apply_sparse(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = super().get_config() config.update({"""weight_decay_rate""": self.weight_decay_rate} ) return config def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Union[str, Any]: if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(lowerCAmelCase_ , lowerCAmelCase_ ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(lowerCAmelCase_ , lowerCAmelCase_ ) is not None: return False return True class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self ) -> Any: _A = [] _A = None @property def UpperCAmelCase ( self ) -> Union[str, Any]: if self._accum_steps is None: _A = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=lowerCAmelCase_ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def UpperCAmelCase ( self ) -> List[str]: if not self._gradients: raise ValueError("""The accumulator should be called first to initialize the gradients""" ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self , lowerCAmelCase_ ) -> List[str]: if not self._gradients: _A = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(lowerCAmelCase_ ) , trainable=lowerCAmelCase_ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(lowerCAmelCase_ ) != len(self._gradients ): raise ValueError(F'''Expected {len(self._gradients )} gradients, but got {len(lowerCAmelCase_ )}''' ) for accum_gradient, gradient in zip(self._gradients , lowerCAmelCase_ ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(lowerCAmelCase_ ) self._accum_steps.assign_add(1 ) def UpperCAmelCase ( self ) -> int: if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(lowerCAmelCase_ ) )
83
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _SCREAMING_SNAKE_CASE = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def snake_case ( snake_case__ :Union[str, Any]) -> Dict: _A = torch.load(snake_case__ , map_location="""cpu""") return sd def snake_case ( snake_case__ :List[str] , snake_case__ :Optional[Any] , snake_case__ :int=rename_keys_prefix) -> Optional[Any]: _A = OrderedDict() _A = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _A = key for name_pair in rename_keys_prefix: _A = new_key.replace(name_pair[0] , name_pair[1]) _A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _A = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def snake_case ( snake_case__ :Tuple , snake_case__ :Tuple) -> int: assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _A = """pretraining""" if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} elif "nlvr" in checkpoint_path: _A = {"""visual_embedding_dim""": 1_024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: _A = {"""visual_embedding_dim""": 512} _A = """multichoice""" elif "vqa_advanced" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048} _A = """vqa_advanced""" elif "vqa" in checkpoint_path: _A = {"""visual_embedding_dim""": 2_048, """num_labels""": 3_129} _A = """vqa""" elif "nlvr" in checkpoint_path: _A = { """visual_embedding_dim""": 1_024, """num_labels""": 2, } _A = """nlvr""" _A = VisualBertConfig(**snake_case__) # Load State Dict _A = load_state_dict(snake_case__) _A = get_new_dict(snake_case__ , snake_case__) if model_type == "pretraining": _A = VisualBertForPreTraining(snake_case__) elif model_type == "vqa": _A = VisualBertForQuestionAnswering(snake_case__) elif model_type == "nlvr": _A = VisualBertForVisualReasoning(snake_case__) elif model_type == "multichoice": _A = VisualBertForMultipleChoice(snake_case__) model.load_state_dict(snake_case__) # Save Checkpoints Path(snake_case__).mkdir(exist_ok=snake_case__) model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
83
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=32 , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=[10, 20, 30, 40] , lowerCAmelCase_=[2, 2, 3, 2] , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=10 , lowerCAmelCase_=0.02 , lowerCAmelCase_=["stage2", "stage3", "stage4"] , lowerCAmelCase_=3 , lowerCAmelCase_=None , ) -> Tuple: _A = parent _A = batch_size _A = image_size _A = num_channels _A = num_stages _A = hidden_sizes _A = depths _A = is_training _A = use_labels _A = intermediate_size _A = hidden_act _A = type_sequence_label_size _A = initializer_range _A = out_features _A = num_labels _A = scope _A = num_stages def UpperCAmelCase ( self ) -> str: _A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self ) -> Any: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def UpperCAmelCase ( self ) -> int: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=lowerCAmelCase_ , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=lowerCAmelCase_ , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> int: _A = UperNetForSemanticSegmentation(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCamelCase :Optional[Any] = {'''image-segmentation''': UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCamelCase :Optional[Any] = False lowerCamelCase :str = False lowerCamelCase :int = False lowerCamelCase :Tuple = False lowerCamelCase :Dict = False lowerCamelCase :Optional[Any] = False def UpperCAmelCase ( self ) -> Union[str, Any]: _A = UperNetModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self ) -> Any: return def UpperCAmelCase ( self ) -> Any: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(lowerCAmelCase_ ) _A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A = [*signature.parameters.keys()] _A = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase_ ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def UpperCAmelCase ( self ) -> Any: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def UpperCAmelCase ( self ) -> str: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def UpperCAmelCase ( self ) -> List[str]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def UpperCAmelCase ( self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCAmelCase ( self ) -> Optional[int]: pass def UpperCAmelCase ( self ) -> str: def check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): _A = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) ) _A = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _A = self.model_tester.num_stages self.assertEqual(len(lowerCAmelCase_ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _A = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[str]: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() _A = _config_zero_init(lowerCAmelCase_ ) _A = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _A = model_class(config=lowerCAmelCase_ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def UpperCAmelCase ( self ) -> List[str]: pass @slow def UpperCAmelCase ( self ) -> Union[str, Any]: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = UperNetForSemanticSegmentation.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def snake_case ( ) -> Tuple: _A = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""") _A = Image.open(snake_case__).convert("""RGB""") return image @require_torch @require_vision @slow class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> str: _A = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _A = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(lowerCAmelCase_ ) _A = prepare_img() _A = processor(images=lowerCAmelCase_ , return_tensors="""pt""" ).to(lowerCAmelCase_ ) with torch.no_grad(): _A = model(**lowerCAmelCase_ ) _A = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) _A = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , lowerCAmelCase_ , atol=1E-4 ) ) def UpperCAmelCase ( self ) -> str: _A = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _A = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(lowerCAmelCase_ ) _A = prepare_img() _A = processor(images=lowerCAmelCase_ , return_tensors="""pt""" ).to(lowerCAmelCase_ ) with torch.no_grad(): _A = model(**lowerCAmelCase_ ) _A = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) _A = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , lowerCAmelCase_ , atol=1E-4 ) )
83
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def UpperCAmelCase ( self ) -> Optional[int]: _A = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCAmelCase_ ): self.assertDictEqual(lowerCAmelCase_ , example_records[i] ) def UpperCAmelCase ( self ) -> str: _A = self._create_example_records() _A = Dataset.from_list(lowerCAmelCase_ ) _A = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def UpperCAmelCase ( self ) -> Any: # checks what happens with missing columns _A = [{"""col_1""": 1}, {"""col_2""": """x"""}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def UpperCAmelCase ( self ) -> Tuple: # checks if the type can be inferred from the second record _A = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _A = Dataset.from_list(lowerCAmelCase_ ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def UpperCAmelCase ( self ) -> Any: _A = Dataset.from_list([] ) self.assertEqual(len(lowerCAmelCase_ ) , 0 ) self.assertListEqual(dset.column_names , [] )
83
1
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case__ , snake_case__))) _A = [float(snake_case__) for n in range(limit + 1)] for p in primes: for n in range(snake_case__ , limit + 1 , snake_case__): phi[n] *= 1 - 1 / p return int(sum(phi[2:])) if __name__ == "__main__": print(F'''{solution() = }''')
83
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case__ , snake_case__))) _A = [float(snake_case__) for n in range(limit + 1)] for p in primes: for n in range(snake_case__ , limit + 1 , snake_case__): phi[n] *= 1 - 1 / p return int(sum(phi[2:])) if __name__ == "__main__": print(F'''{solution() = }''')
83
1
_SCREAMING_SNAKE_CASE = 0 # The first color of the flag. _SCREAMING_SNAKE_CASE = 1 # The second color of the flag. _SCREAMING_SNAKE_CASE = 2 # The third color of the flag. _SCREAMING_SNAKE_CASE = (red, white, blue) def snake_case ( snake_case__ :list) -> list: if not sequence: return [] if len(snake_case__) == 1: return list(snake_case__) _A = 0 _A = len(snake_case__) - 1 _A = 0 while mid <= high: if sequence[mid] == colors[0]: _A , _A = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: _A , _A = sequence[high], sequence[mid] high -= 1 else: _A = F'''The elements inside the sequence must contains only {colors} values''' raise ValueError(snake_case__) return sequence if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = input('Enter numbers separated by commas:\n').strip() _SCREAMING_SNAKE_CASE = [int(item.strip()) for item in user_input.split(',')] print(F'''{dutch_national_flag_sort(unsorted)}''')
83
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=13 , lowerCAmelCase_=7 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=99 , lowerCAmelCase_=32 , lowerCAmelCase_=5 , lowerCAmelCase_=4 , lowerCAmelCase_=37 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=16 , lowerCAmelCase_=2 , lowerCAmelCase_=0.02 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_="None" , lowerCAmelCase_=3 , lowerCAmelCase_=4 , lowerCAmelCase_=None , ) -> Union[str, Any]: _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = num_choices _A = relative_attention _A = position_biased_input _A = pos_att_type _A = scope def UpperCAmelCase ( self ) -> Dict: _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self ) -> Optional[int]: return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Any: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: _A = DebertaVaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ )[0] _A = model(lowerCAmelCase_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> List[str]: _A = DebertaVaForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: _A = self.num_labels _A = DebertaVaForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = self.num_labels _A = DebertaVaForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _A = DebertaVaForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: _A = DebertaVaForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase :int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) lowerCamelCase :str = ( { '''feature-extraction''': DebertaVaModel, '''fill-mask''': DebertaVaForMaskedLM, '''question-answering''': DebertaVaForQuestionAnswering, '''text-classification''': DebertaVaForSequenceClassification, '''token-classification''': DebertaVaForTokenClassification, '''zero-shot''': DebertaVaForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase :str = True lowerCamelCase :Union[str, Any] = False lowerCamelCase :Optional[int] = False lowerCamelCase :List[str] = False lowerCamelCase :str = False def UpperCAmelCase ( self ) -> Optional[int]: _A = DebertaVaModelTester(self ) _A = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> List[str]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> List[str]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Dict: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*lowerCAmelCase_ ) @slow def UpperCAmelCase ( self ) -> Any: for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = DebertaVaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason="""Model not available yet""" ) def UpperCAmelCase ( self ) -> int: pass @slow def UpperCAmelCase ( self ) -> Optional[Any]: _A = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _A = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) _A = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _A = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] # compare the actual values for a slice. _A = torch.tensor( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
83
1
from __future__ import annotations from cmath import sqrt def snake_case ( snake_case__ :int , snake_case__ :int , snake_case__ :int) -> tuple[complex, complex]: if a == 0: raise ValueError("""Coefficient 'a' must not be zero.""") _A = b * b - 4 * a * c _A = (-b + sqrt(snake_case__)) / (2 * a) _A = (-b - sqrt(snake_case__)) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def snake_case ( ) -> List[Any]: _A , _A = quadratic_roots(a=5 , b=6 , c=1) print(F'''The solutions are: {solutiona} and {solutiona}''') if __name__ == "__main__": main()
83
def snake_case ( snake_case__ :int , snake_case__ :int) -> int: return int(input_a == input_a == 0) def snake_case ( ) -> None: print("""Truth Table of NOR Gate:""") print("""| Input 1 | Input 2 | Output |""") print(F'''| 0 | 0 | {nor_gate(0 , 0)} |''') print(F'''| 0 | 1 | {nor_gate(0 , 1)} |''') print(F'''| 1 | 0 | {nor_gate(1 , 0)} |''') print(F'''| 1 | 1 | {nor_gate(1 , 1)} |''') if __name__ == "__main__": import doctest doctest.testmod() main()
83
1
def snake_case ( snake_case__ :dict) -> set: _A = set() # edges = list of graph's edges _A = get_edges(snake_case__) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: _A , _A = edges.pop() chosen_vertices.add(snake_case__) chosen_vertices.add(snake_case__) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(snake_case__) return chosen_vertices def snake_case ( snake_case__ :dict) -> set: _A = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node)) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
83
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=sys.maxsize ) -> str: _A = """bilinear""" _A = max_size _A = short_edge_length def __call__( self , lowerCAmelCase_ ) -> Optional[Any]: _A = [] for img in imgs: _A , _A = img.shape[:2] # later: provide list and randomly choose index for resize _A = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img _A = size * 1.0 / min(lowerCAmelCase_ , lowerCAmelCase_ ) if h < w: _A , _A = size, scale * w else: _A , _A = scale * h, size if max(lowerCAmelCase_ , lowerCAmelCase_ ) > self.max_size: _A = self.max_size * 1.0 / max(lowerCAmelCase_ , lowerCAmelCase_ ) _A = newh * scale _A = neww * scale _A = int(neww + 0.5 ) _A = int(newh + 0.5 ) if img.dtype == np.uinta: _A = Image.fromarray(lowerCAmelCase_ ) _A = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) _A = np.asarray(lowerCAmelCase_ ) else: _A = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw _A = nn.functional.interpolate( lowerCAmelCase_ , (newh, neww) , mode=self.interp_method , align_corners=lowerCAmelCase_ ).squeeze(0 ) img_augs.append(lowerCAmelCase_ ) return img_augs class a : """simple docstring""" def __init__( self , lowerCAmelCase_ ) -> List[Any]: _A = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) _A = cfg.INPUT.FORMAT _A = cfg.SIZE_DIVISIBILITY _A = cfg.PAD_VALUE _A = cfg.INPUT.MAX_SIZE_TEST _A = cfg.MODEL.DEVICE _A = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) _A = lambda lowerCAmelCase_ : (x - self.pixel_mean) / self.pixel_std def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = tuple(max(lowerCAmelCase_ ) for s in zip(*[img.shape for img in images] ) ) _A = [im.shape[-2:] for im in images] _A = [ nn.functional.pad( lowerCAmelCase_ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] return torch.stack(lowerCAmelCase_ ), torch.tensor(lowerCAmelCase_ ) def __call__( self , lowerCAmelCase_ , lowerCAmelCase_=False ) -> int: with torch.no_grad(): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = [images] if single_image: assert len(lowerCAmelCase_ ) == 1 for i in range(len(lowerCAmelCase_ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(lowerCAmelCase_ , images.pop(lowerCAmelCase_ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( lowerCAmelCase_ , torch.as_tensor(img_tensorize(images.pop(lowerCAmelCase_ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge _A = torch.tensor([im.shape[:2] for im in images] ) _A = self.aug(lowerCAmelCase_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic _A = [self.normalizer(lowerCAmelCase_ ) for x in images] # now pad them to do the following operations _A , _A = self.pad(lowerCAmelCase_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad _A = torch.true_divide(lowerCAmelCase_ , lowerCAmelCase_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[Any]) -> Tuple: boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def snake_case ( snake_case__ :Optional[int] , snake_case__ :Tuple[int, int]) -> Optional[Any]: assert torch.isfinite(snake_case__).all(), "Box tensor contains infinite or NaN!" _A , _A = box_size tensor[:, 0].clamp_(min=0 , max=snake_case__) tensor[:, 1].clamp_(min=0 , max=snake_case__) tensor[:, 2].clamp_(min=0 , max=snake_case__) tensor[:, 3].clamp_(min=0 , max=snake_case__)
83
1
from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''' , __lowerCAmelCase , ) class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :List[str] = RobertaConfig lowerCamelCase :int = '''roberta''' def __init__( self , lowerCAmelCase_ ) -> List[str]: super().__init__(lowerCAmelCase_ ) _A = RobertaEmbeddings(lowerCAmelCase_ ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''' , __lowerCAmelCase , ) class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Optional[int] = RobertaConfig lowerCamelCase :str = '''roberta''' def __init__( self , lowerCAmelCase_ ) -> List[str]: super().__init__(lowerCAmelCase_ ) _A = config.num_labels _A = config.num_hidden_layers _A = DeeRobertaModel(lowerCAmelCase_ ) _A = nn.Dropout(config.hidden_dropout_prob ) _A = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=-1 , lowerCAmelCase_=False , ) -> List[str]: _A = self.num_layers try: _A = self.roberta( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , position_ids=lowerCAmelCase_ , head_mask=lowerCAmelCase_ , inputs_embeds=lowerCAmelCase_ , ) _A = outputs[1] _A = self.dropout(lowerCAmelCase_ ) _A = self.classifier(lowerCAmelCase_ ) _A = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: _A = e.message _A = e.exit_layer _A = outputs[0] if not self.training: _A = entropy(lowerCAmelCase_ ) _A = [] _A = [] if labels is not None: if self.num_labels == 1: # We are doing regression _A = MSELoss() _A = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: _A = CrossEntropyLoss() _A = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits _A = [] for highway_exit in outputs[-1]: _A = highway_exit[0] if not self.training: highway_logits_all.append(lowerCAmelCase_ ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression _A = MSELoss() _A = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: _A = CrossEntropyLoss() _A = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(lowerCAmelCase_ ) if train_highway: _A = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: _A = (loss,) + outputs if not self.training: _A = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: _A = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
83
from collections import defaultdict def snake_case ( snake_case__ :int) -> int: _A = 1 _A = True for v in tree[start]: if v not in visited: ret += dfs(snake_case__) if ret % 2 == 0: cuts.append(snake_case__) return ret def snake_case ( ) -> Any: dfs(1) if __name__ == "__main__": _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10, 9 _SCREAMING_SNAKE_CASE = defaultdict(list) _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
83
1
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( '--original_config_file', default=None, type=str, help='The YAML config file corresponding to the original architecture.', ) parser.add_argument( '--num_in_channels', default=None, type=int, help='The number of input channels. If `None` number of input channels will be automatically inferred.', ) parser.add_argument( '--scheduler_type', default='pndm', type=str, help='Type of scheduler to use. Should be one of [\'pndm\', \'lms\', \'ddim\', \'euler\', \'euler-ancestral\', \'dpm\']', ) parser.add_argument( '--pipeline_type', default=None, type=str, help=( 'The pipeline type. One of \'FrozenOpenCLIPEmbedder\', \'FrozenCLIPEmbedder\', \'PaintByExample\'' '. If `None` pipeline will be automatically inferred.' ), ) parser.add_argument( '--image_size', default=None, type=int, help=( 'The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2' ' Base. Use 768 for Stable Diffusion v2.' ), ) parser.add_argument( '--prediction_type', default=None, type=str, help=( 'The prediction type that the model was trained on. Use \'epsilon\' for Stable Diffusion v1.X and Stable' ' Diffusion v2 Base. Use \'v_prediction\' for Stable Diffusion v2.' ), ) parser.add_argument( '--extract_ema', action='store_true', help=( 'Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights' ' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield' ' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.' ), ) parser.add_argument( '--upcast_attention', action='store_true', help=( 'Whether the attention computation should always be upcasted. This is necessary when running stable' ' diffusion 2.1.' ), ) parser.add_argument( '--from_safetensors', action='store_true', help='If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.', ) parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') parser.add_argument( '--stable_unclip', type=str, default=None, required=False, help='Set if this is a stable unCLIP model. One of \'txt2img\' or \'img2img\'.', ) parser.add_argument( '--stable_unclip_prior', type=str, default=None, required=False, help='Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.', ) parser.add_argument( '--clip_stats_path', type=str, help='Path to the clip stats file. Only required if the stable unclip model\'s config specifies `model.params.noise_aug_config.params.clip_stats_path`.', required=False, ) parser.add_argument( '--controlnet', action='store_true', default=None, help='Set flag if this is a controlnet checkpoint.' ) parser.add_argument('--half', action='store_true', help='Save weights in half precision.') parser.add_argument( '--vae_path', type=str, default=None, required=False, help='Set to a path, hub id to an already converted vae to not convert it again.', ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
83
import heapq def snake_case ( snake_case__ :dict) -> set[int]: _A = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(snake_case__ , [-1 * len(snake_case__), (key, value)]) # chosen_vertices = set of chosen vertices _A = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _A = heapq.heappop(snake_case__)[1][0] chosen_vertices.add(snake_case__) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _A = elem[1][1].index(snake_case__) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(snake_case__) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(F'''Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}''')
83
1
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase_=0.01 , lowerCAmelCase_=10_00 ) -> Tuple: _A = p_stop _A = max_length def __iter__( self ) -> Tuple: _A = 0 _A = False while not stop and count < self.max_length: yield count count += 1 _A = random.random() < self.p_stop class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=True ) -> Union[str, Any]: _A = [ BatchSamplerShard(lowerCAmelCase_ , 2 , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) for i in range(2 ) ] _A = [list(lowerCAmelCase_ ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowerCAmelCase_ ) for shard in batch_sampler_shards] , [len(lowerCAmelCase_ ) for e in expected] ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[str]: # Check the shards when the dataset is a round multiple of total batch size. _A = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) _A = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. _A = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) _A = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. _A = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) _A = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. _A = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) _A = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) # Check the shards when the dataset is very small. _A = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) _A = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: # Check the shards when the dataset is a round multiple of batch size. _A = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) _A = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. _A = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) _A = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. _A = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) _A = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. _A = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) _A = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> str: # Check the shards when the dataset is a round multiple of total batch size. _A = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. _A = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. _A = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. _A = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. _A = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCAmelCase_ ) _A = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[Any]: # Check the shards when the dataset is a round multiple of batch size. _A = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. _A = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. _A = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. _A = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) _A = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_ , lowerCAmelCase_ , split_batches=lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Tuple: _A = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] _A = [BatchSamplerShard(lowerCAmelCase_ , 2 , lowerCAmelCase_ , even_batches=lowerCAmelCase_ ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=2 , lowerCAmelCase_=False ) -> int: random.seed(lowerCAmelCase_ ) _A = list(lowerCAmelCase_ ) _A = [ IterableDatasetShard( lowerCAmelCase_ , batch_size=lowerCAmelCase_ , drop_last=lowerCAmelCase_ , num_processes=lowerCAmelCase_ , process_index=lowerCAmelCase_ , split_batches=lowerCAmelCase_ , ) for i in range(lowerCAmelCase_ ) ] _A = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowerCAmelCase_ ) iterable_dataset_lists.append(list(lowerCAmelCase_ ) ) _A = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size _A = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowerCAmelCase_ ) , len(lowerCAmelCase_ ) ) self.assertTrue(len(lowerCAmelCase_ ) % shard_batch_size == 0 ) _A = [] for idx in range(0 , len(lowerCAmelCase_ ) , lowerCAmelCase_ ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowerCAmelCase_ ) < len(lowerCAmelCase_ ): reference += reference self.assertListEqual(lowerCAmelCase_ , reference[: len(lowerCAmelCase_ )] ) def UpperCAmelCase ( self ) -> List[str]: _A = 42 _A = RandomIterableDataset() self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) # Edge case with a very small dataset _A = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_ , lowerCAmelCase_ , batch_size=4 , drop_last=lowerCAmelCase_ , split_batches=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: _A = BatchSampler(range(16 ) , batch_size=4 , drop_last=lowerCAmelCase_ ) _A = SkipBatchSampler(lowerCAmelCase_ , 2 ) self.assertListEqual(list(lowerCAmelCase_ ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCAmelCase ( self ) -> Dict: _A = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCAmelCase ( self ) -> Optional[int]: _A = DataLoader(list(range(16 ) ) , batch_size=4 ) _A = skip_first_batches(lowerCAmelCase_ , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCAmelCase ( self ) -> Any: _A = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def UpperCAmelCase ( self ) -> Dict: Accelerator() _A = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
83
import math import unittest def snake_case ( snake_case__ :int) -> bool: assert isinstance(snake_case__ , snake_case__) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def UpperCAmelCase ( self ) -> Dict: with self.assertRaises(lowerCAmelCase_ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , """Zero doesn't have any positive factors, primes must have exactly two.""" , ) self.assertFalse( is_prime(1 ) , """One only has 1 positive factor, primes must have exactly two.""" , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
83
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _SCREAMING_SNAKE_CASE = { 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConfig', 'SqueezeBertOnnxConfig', ], 'tokenization_squeezebert': ['SqueezeBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['SqueezeBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'SqueezeBertForMaskedLM', 'SqueezeBertForMultipleChoice', 'SqueezeBertForQuestionAnswering', 'SqueezeBertForSequenceClassification', 'SqueezeBertForTokenClassification', 'SqueezeBertModel', 'SqueezeBertModule', 'SqueezeBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
1
from __future__ import annotations import math def snake_case ( snake_case__ :float , snake_case__ :int) -> float: _A = u for i in range(1 , snake_case__): _A = temp * (u - i) return temp def snake_case ( ) -> None: _A = int(input("""enter the numbers of values: """)) _A = [] for _ in range(snake_case__): y.append([]) for i in range(snake_case__): for j in range(snake_case__): y[i].append(snake_case__) _A = 0 print("""enter the values of parameters in a list: """) _A = list(map(snake_case__ , input().split())) print("""enter the values of corresponding parameters: """) for i in range(snake_case__): _A = float(input()) _A = int(input("""enter the value to interpolate: """)) _A = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , snake_case__): for j in range(n - i): _A = y[j + 1][i - 1] - y[j][i - 1] _A = y[0][0] for i in range(1 , snake_case__): summ += (ucal(snake_case__ , snake_case__) * y[0][i]) / math.factorial(snake_case__) print(F'''the value at {value} is {summ}''') if __name__ == "__main__": main()
83
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(__lowerCAmelCase ) class a ( __lowerCAmelCase ): """simple docstring""" def __init__( self , *lowerCAmelCase_ , **lowerCAmelCase_ ) -> Optional[Any]: super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) self.check_model_type(lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , **lowerCAmelCase_ ) -> Tuple: _A , _A = {}, {} if padding is not None: _A = padding if truncation is not None: _A = truncation if top_k is not None: _A = top_k return preprocess_params, {}, postprocess_params def __call__( self , lowerCAmelCase_ , lowerCAmelCase_ = None , **lowerCAmelCase_ ) -> Union[str, Any]: if isinstance(lowerCAmelCase_ , (Image.Image, str) ) and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = {"""image""": image, """question""": question} else: _A = image _A = super().__call__(lowerCAmelCase_ , **lowerCAmelCase_ ) return results def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False ) -> Any: _A = load_image(inputs["""image"""] ) _A = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ ) _A = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase_ ) return model_inputs def UpperCAmelCase ( self , lowerCAmelCase_ ) -> Tuple: _A = self.model(**lowerCAmelCase_ ) return model_outputs def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=5 ) -> Union[str, Any]: if top_k > self.model.config.num_labels: _A = self.model.config.num_labels if self.framework == "pt": _A = model_outputs.logits.sigmoid()[0] _A , _A = probs.topk(lowerCAmelCase_ ) else: raise ValueError(F'''Unsupported framework: {self.framework}''' ) _A = scores.tolist() _A = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowerCAmelCase_ , lowerCAmelCase_ )]
83
1
from __future__ import annotations def snake_case ( snake_case__ :list , snake_case__ :int , snake_case__ :int , snake_case__ :int) -> list: _A = [] _A , _A = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0)) _A = result + left + right return input_list def snake_case ( snake_case__ :list) -> list: if len(snake_case__) <= 1: return input_list _A = list(snake_case__) # iteration for two-way merging _A = 2 while p <= len(snake_case__): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(snake_case__) , snake_case__): _A = i _A = i + p - 1 _A = (low + high + 1) // 2 _A = merge(snake_case__ , snake_case__ , snake_case__ , snake_case__) # final merge of last two parts if p * 2 >= len(snake_case__): _A = i _A = merge(snake_case__ , 0 , snake_case__ , len(snake_case__) - 1) break p *= 2 return input_list if __name__ == "__main__": _SCREAMING_SNAKE_CASE = input('Enter numbers separated by a comma:\n').strip() if user_input == "": _SCREAMING_SNAKE_CASE = [] else: _SCREAMING_SNAKE_CASE = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
83
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :str , snake_case__ :PreTrainedTokenizer , snake_case__ :int , snake_case__ :Optional[int] = None , ) -> Optional[int]: _A = {} if train_file is not None: _A = [train_file] if eval_file is not None: _A = [eval_file] if test_file is not None: _A = [test_file] _A = datasets.load_dataset("""csv""" , data_files=snake_case__) _A = list(ds[list(files.keys())[0]].features.keys()) _A = features_name.pop(snake_case__) _A = list(set(ds[list(files.keys())[0]][label_name])) _A = {label: i for i, label in enumerate(snake_case__)} _A = tokenizer.model_input_names _A = {} if len(snake_case__) == 1: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( example[features_name[0]] , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""") , batched=snake_case__ , ) elif len(snake_case__) == 2: for k in files.keys(): _A = ds[k].map( lambda snake_case__: tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=snake_case__ , max_length=snake_case__ , padding="""max_length""" , ) , batched=snake_case__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _A = {k: v for k, v in ex.items() if k in input_names} _A = labelaid[ex[label_name]] yield (d, label) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _A = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _A = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION]))) _A = ( tf.data.Dataset.from_generator( snake_case__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None]) for k in input_names}, tf.TensorShape([])) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _A = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST]))) return train_ds, val_ds, test_ds, labelaid _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class a : """simple docstring""" lowerCamelCase :int = field(metadata={'''help''': '''Which column contains the label'''} ) lowerCamelCase :str = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the training file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the development file'''} ) lowerCamelCase :Optional[str] = field(default=__lowerCAmelCase , metadata={'''help''': '''The path of the test file'''} ) lowerCamelCase :int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) lowerCamelCase :bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) @dataclass class a : """simple docstring""" lowerCamelCase :str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) lowerCamelCase :bool = field(default=__lowerCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. lowerCamelCase :Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) def snake_case ( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments)) _A , _A , _A = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir) and os.listdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""") # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info( F'''n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1)}, ''' F'''16-bits training: {training_args.fpaa}''') logger.info(F'''Training/evaluation parameters {training_args}''') # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _A = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _A , _A , _A , _A = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=snake_case__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _A = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(snake_case__) , labelaid=snake_case__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="""text-classification""" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _A = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(""".bin""" in model_args.model_name_or_path) , config=snake_case__ , cache_dir=model_args.cache_dir , ) def compute_metrics(snake_case__ :EvalPrediction) -> Dict: _A = np.argmax(p.predictions , axis=1) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _A = TFTrainer( model=snake_case__ , args=snake_case__ , train_dataset=snake_case__ , eval_dataset=snake_case__ , compute_metrics=snake_case__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir) # Evaluation _A = {} if training_args.do_eval: logger.info("""*** Evaluate ***""") _A = trainer.evaluate() _A = os.path.join(training_args.output_dir , """eval_results.txt""") with open(snake_case__ , """w""") as writer: logger.info("""***** Eval results *****""") for key, value in result.items(): logger.info(F''' {key} = {value}''') writer.write(F'''{key} = {value}\n''') results.update(snake_case__) return results if __name__ == "__main__": main()
83
1
_SCREAMING_SNAKE_CASE = { 0: '0', 1: '1', 2: '2', 3: '3', 4: '4', 5: '5', 6: '6', 7: '7', 8: '8', 9: '9', 10: 'a', 11: 'b', 12: 'c', 13: 'd', 14: 'e', 15: 'f', } def snake_case ( snake_case__ :float) -> str: assert type(snake_case__) in (int, float) and decimal == int(snake_case__) _A = int(snake_case__) _A = """""" _A = False if decimal < 0: _A = True decimal *= -1 while decimal > 0: _A , _A = divmod(snake_case__ , 16) _A = values[remainder] + hexadecimal _A = """0x""" + hexadecimal if negative: _A = """-""" + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
83
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Union[str, Any] = '''speech_to_text''' lowerCamelCase :List[str] = ['''past_key_values'''] lowerCamelCase :str = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase_=1_00_00 , lowerCAmelCase_=12 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=6 , lowerCAmelCase_=20_48 , lowerCAmelCase_=4 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_="relu" , lowerCAmelCase_=2_56 , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.02 , lowerCAmelCase_=2 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=0 , lowerCAmelCase_=2 , lowerCAmelCase_=60_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=2 , lowerCAmelCase_=(5, 5) , lowerCAmelCase_=10_24 , lowerCAmelCase_=80 , lowerCAmelCase_=1 , **lowerCAmelCase_ , ) -> Tuple: _A = vocab_size _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = encoder_layerdrop _A = decoder_layerdrop _A = use_cache _A = encoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = max_source_positions _A = max_target_positions _A = num_conv_layers _A = list(lowerCAmelCase_ ) _A = conv_channels _A = input_feat_per_channel _A = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` """ F'''but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, ''' F'''`config.num_conv_layers = {self.num_conv_layers}`.''' ) super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , )
83
1