code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from math import factorial _a = {str(d): factorial(d) for d in range(10)} def _A ( UpperCamelCase_ : int) -> int: '''simple docstring''' return sum(DIGIT_FACTORIAL[d] for d in str(UpperCamelCase_)) def _A ( ) -> int: '''simple docstring''' __lowercase = 7 * factorial(9) + 1 return sum(i for i in range(3, UpperCamelCase_) if sum_of_digit_factorial(UpperCamelCase_) == i) if __name__ == "__main__": print(F"{solution() = }")
17
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class _lowerCAmelCase ( unittest.TestCase ,lowercase ): """simple docstring""" def _lowercase ( self : List[Any] ): __lowercase = load_tool("text-classification" ) self.tool.setup() __lowercase = load_tool("text-classification", remote=UpperCAmelCase__ ) def _lowercase ( self : str ): __lowercase = self.tool("That's quite cool", ["positive", "negative"] ) self.assertEqual(UpperCAmelCase__, "positive" ) def _lowercase ( self : str ): __lowercase = self.remote_tool("That's quite cool", ["positive", "negative"] ) self.assertEqual(UpperCAmelCase__, "positive" ) def _lowercase ( self : List[str] ): __lowercase = self.tool(text="That's quite cool", labels=["positive", "negative"] ) self.assertEqual(UpperCAmelCase__, "positive" ) def _lowercase ( self : Tuple ): __lowercase = self.remote_tool(text="That's quite cool", labels=["positive", "negative"] ) self.assertEqual(UpperCAmelCase__, "positive" )
17
1
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html lowerCAmelCase__ : Any ='''platform''' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def __lowercase ( a__ , a__ , a__=None , a__=None , a__=None , a__=None , a__=None , a__=None , ) -> Dict: if attention_mask is None: __SCREAMING_SNAKE_CASE = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: __SCREAMING_SNAKE_CASE = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: __SCREAMING_SNAKE_CASE = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __SCREAMING_SNAKE_CASE = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __SCREAMING_SNAKE_CASE = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class UpperCAmelCase_ : '''simple docstring''' def __init__( self , _A , _A=13 , _A=7 , _A=True , _A=False , _A=99 , _A=16 , _A=2 , _A=4 , _A=4 , _A="gelu" , _A=0.1 , _A=0.1 , _A=32 , _A=2 , _A=1 , _A=0 , _A=0.0_2 , ): '''simple docstring''' __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = eos_token_id __SCREAMING_SNAKE_CASE = pad_token_id __SCREAMING_SNAKE_CASE = bos_token_id __SCREAMING_SNAKE_CASE = initializer_range def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) __SCREAMING_SNAKE_CASE = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) __SCREAMING_SNAKE_CASE = shift_tokens_right(_A , 1 , 2 ) __SCREAMING_SNAKE_CASE = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=_A , ) __SCREAMING_SNAKE_CASE = prepare_blenderbot_inputs_dict(_A , _A , _A ) return config, inputs_dict def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() return config, inputs_dict def _A ( self , _A , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 20 __SCREAMING_SNAKE_CASE = model_class_name(_A ) __SCREAMING_SNAKE_CASE = model.encode(inputs_dict['input_ids'] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) __SCREAMING_SNAKE_CASE = model.init_cache(decoder_input_ids.shape[0] , _A , _A ) __SCREAMING_SNAKE_CASE = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='i4' ) __SCREAMING_SNAKE_CASE = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) __SCREAMING_SNAKE_CASE = model.decode( decoder_input_ids[:, :-1] , _A , decoder_attention_mask=_A , past_key_values=_A , decoder_position_ids=_A , ) __SCREAMING_SNAKE_CASE = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) __SCREAMING_SNAKE_CASE = model.decode( decoder_input_ids[:, -1:] , _A , decoder_attention_mask=_A , past_key_values=outputs_cache.past_key_values , decoder_position_ids=_A , ) __SCREAMING_SNAKE_CASE = model.decode(_A , _A ) __SCREAMING_SNAKE_CASE = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def _A ( self , _A , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 20 __SCREAMING_SNAKE_CASE = model_class_name(_A ) __SCREAMING_SNAKE_CASE = model.encode(inputs_dict['input_ids'] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) __SCREAMING_SNAKE_CASE = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) __SCREAMING_SNAKE_CASE = model.init_cache(decoder_input_ids.shape[0] , _A , _A ) __SCREAMING_SNAKE_CASE = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) __SCREAMING_SNAKE_CASE = model.decode( decoder_input_ids[:, :-1] , _A , decoder_attention_mask=_A , past_key_values=_A , decoder_position_ids=_A , ) __SCREAMING_SNAKE_CASE = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) __SCREAMING_SNAKE_CASE = model.decode( decoder_input_ids[:, -1:] , _A , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=_A , decoder_position_ids=_A , ) __SCREAMING_SNAKE_CASE = model.decode(_A , _A , decoder_attention_mask=_A ) __SCREAMING_SNAKE_CASE = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : List[Any] = 99 def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) __SCREAMING_SNAKE_CASE = input_ids.shape[0] __SCREAMING_SNAKE_CASE = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self._get_config_and_data() __SCREAMING_SNAKE_CASE = FlaxBlenderbotSmallForConditionalGeneration(_A ) __SCREAMING_SNAKE_CASE = lm_model(input_ids=_A ) __SCREAMING_SNAKE_CASE = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs['logits'].shape , _A ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) __SCREAMING_SNAKE_CASE = FlaxBlenderbotSmallForConditionalGeneration(_A ) __SCREAMING_SNAKE_CASE = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) __SCREAMING_SNAKE_CASE = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) __SCREAMING_SNAKE_CASE = lm_model(input_ids=_A , decoder_input_ids=_A ) __SCREAMING_SNAKE_CASE = (*summary.shape, config.vocab_size) self.assertEqual(outputs['logits'].shape , _A ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) __SCREAMING_SNAKE_CASE = shift_tokens_right(_A , 1 , 2 ) __SCREAMING_SNAKE_CASE = np.equal(_A , 1 ).astype(np.floataa ).sum() __SCREAMING_SNAKE_CASE = np.equal(_A , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(_A , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class UpperCAmelCase_ ( UpperCamelCase_ , unittest.TestCase , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ : Optional[int] = True UpperCamelCase__ : List[str] = ( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) UpperCamelCase__ : Dict = (FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = FlaxBlenderbotSmallModelTester(self ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(_A , _A , _A ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(_A , _A , _A ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __SCREAMING_SNAKE_CASE = self._prepare_for_class(_A , _A ) __SCREAMING_SNAKE_CASE = model_class(_A ) @jax.jit def encode_jitted(_A , _A=None , **_A ): return model.encode(input_ids=_A , attention_mask=_A ) with self.subTest('JIT Enabled' ): __SCREAMING_SNAKE_CASE = encode_jitted(**_A ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): __SCREAMING_SNAKE_CASE = encode_jitted(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) ) for jitted_output, output in zip(_A , _A ): self.assertEqual(jitted_output.shape , output.shape ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __SCREAMING_SNAKE_CASE = model_class(_A ) __SCREAMING_SNAKE_CASE = model.encode(inputs_dict['input_ids'] , inputs_dict['attention_mask'] ) __SCREAMING_SNAKE_CASE = { 'decoder_input_ids': inputs_dict['decoder_input_ids'], 'decoder_attention_mask': inputs_dict['decoder_attention_mask'], 'encoder_outputs': encoder_outputs, } @jax.jit def decode_jitted(_A , _A , _A ): return model.decode( decoder_input_ids=_A , decoder_attention_mask=_A , encoder_outputs=_A , ) with self.subTest('JIT Enabled' ): __SCREAMING_SNAKE_CASE = decode_jitted(**_A ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): __SCREAMING_SNAKE_CASE = decode_jitted(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) ) for jitted_output, output in zip(_A , _A ): self.assertEqual(jitted_output.shape , output.shape ) @slow def _A ( self ): '''simple docstring''' for model_class_name in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class_name.from_pretrained('facebook/blenderbot_small-90M' ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids __SCREAMING_SNAKE_CASE = np.ones((1, 1) ) * model.config.eos_token_id __SCREAMING_SNAKE_CASE = model(_A ) self.assertIsNotNone(_A )
354
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __lowercase ( a__ ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] __SCREAMING_SNAKE_CASE = [5, 5, 5, 5] elif "fl4" in model_name: __SCREAMING_SNAKE_CASE = [4, 4, 4, 4] __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] if "lrf" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] else: __SCREAMING_SNAKE_CASE = [2, 2, 2, 2] if "tiny" in model_name: __SCREAMING_SNAKE_CASE = 96 elif "small" in model_name: __SCREAMING_SNAKE_CASE = 96 elif "base" in model_name: __SCREAMING_SNAKE_CASE = 1_28 elif "large" in model_name: __SCREAMING_SNAKE_CASE = 1_92 elif "xlarge" in model_name: __SCREAMING_SNAKE_CASE = 2_56 elif "huge" in model_name: __SCREAMING_SNAKE_CASE = 3_52 # set label information __SCREAMING_SNAKE_CASE = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: __SCREAMING_SNAKE_CASE = 'imagenet-22k-id2label.json' else: __SCREAMING_SNAKE_CASE = 'imagenet-1k-id2label.json' __SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(a__ , a__ , repo_type='dataset' ) , 'r' ) ) __SCREAMING_SNAKE_CASE = {int(a__ ): v for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = FocalNetConfig( embed_dim=a__ , depths=a__ , focal_levels=a__ , focal_windows=a__ , use_conv_embed=a__ , idalabel=a__ , labelaid=a__ , use_post_layernorm=a__ , use_layerscale=a__ , ) return config def __lowercase ( a__ ) -> Any: if "patch_embed.proj" in name: __SCREAMING_SNAKE_CASE = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: __SCREAMING_SNAKE_CASE = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: __SCREAMING_SNAKE_CASE = 'encoder.' + name if "encoder.layers" in name: __SCREAMING_SNAKE_CASE = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: __SCREAMING_SNAKE_CASE = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: __SCREAMING_SNAKE_CASE = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": __SCREAMING_SNAKE_CASE = 'layernorm.weight' if name == "norm.bias": __SCREAMING_SNAKE_CASE = 'layernorm.bias' if "head" in name: __SCREAMING_SNAKE_CASE = name.replace('head' , 'classifier' ) else: __SCREAMING_SNAKE_CASE = 'focalnet.' + name return name def __lowercase ( a__ , a__ , a__=False ) -> Dict: # fmt: off __SCREAMING_SNAKE_CASE = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on __SCREAMING_SNAKE_CASE = model_name_to_url[model_name] print('Checkpoint URL: ' , a__ ) __SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(a__ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): __SCREAMING_SNAKE_CASE = state_dict.pop(a__ ) __SCREAMING_SNAKE_CASE = val __SCREAMING_SNAKE_CASE = get_focalnet_config(a__ ) __SCREAMING_SNAKE_CASE = FocalNetForImageClassification(a__ ) model.eval() # load state dict model.load_state_dict(a__ ) # verify conversion __SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' __SCREAMING_SNAKE_CASE = BitImageProcessor( do_resize=a__ , size={'shortest_edge': 2_56} , resample=PILImageResampling.BILINEAR , do_center_crop=a__ , crop_size=2_24 , do_normalize=a__ , image_mean=a__ , image_std=a__ , ) __SCREAMING_SNAKE_CASE = Image.open(requests.get(a__ , stream=a__ ).raw ) __SCREAMING_SNAKE_CASE = processor(images=a__ , return_tensors='pt' ) __SCREAMING_SNAKE_CASE = transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) __SCREAMING_SNAKE_CASE = image_transforms(a__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , a__ , atol=1E-4 ) __SCREAMING_SNAKE_CASE = model(**a__ ) __SCREAMING_SNAKE_CASE = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": __SCREAMING_SNAKE_CASE = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": __SCREAMING_SNAKE_CASE = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": __SCREAMING_SNAKE_CASE = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] , a__ , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and processor of {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a__ ) processor.save_pretrained(a__ ) if push_to_hub: print(f"""Pushing model and processor of {model_name} to the hub...""" ) model.push_to_hub(f"""{model_name}""" ) processor.push_to_hub(f"""{model_name}""" ) if __name__ == "__main__": lowerCAmelCase__ : int =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''focalnet-tiny''', type=str, help='''Name of the FocalNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub.''', ) lowerCAmelCase__ : List[Any] =parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
118
0
'''simple docstring''' import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class a_ ( unittest.TestCase ): '''simple docstring''' def snake_case_( self ) -> int: _SCREAMING_SNAKE_CASE = inspect.getfile(accelerate.test_utils ) _SCREAMING_SNAKE_CASE = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """external_deps""", """test_metrics.py"""] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 _SCREAMING_SNAKE_CASE = test_metrics @require_cpu def snake_case_( self ) -> int: debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def snake_case_( self ) -> str: debug_launcher(self.test_metrics.main ) @require_single_gpu def snake_case_( self ) -> Optional[Any]: self.test_metrics.main() @require_multi_gpu def snake_case_( self ) -> str: print(f'Found {torch.cuda.device_count()} devices.' ) _SCREAMING_SNAKE_CASE = ["""torchrun""", f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(A , env=os.environ.copy() )
58
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a_ = logging.get_logger(__name__) @dataclass class UpperCAmelCase_ ( snake_case ): UpperCamelCase =[ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self , **UpperCamelCase_ ) -> Optional[Any]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __lowercase : Union[str, Any] = deprecated_arg[3:] setattr(self , UpperCamelCase_ , not kwargs.pop(UpperCamelCase_ ) ) logger.warning( F"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" F""" {positive_arg}={kwargs[positive_arg]}""" ) __lowercase : Dict = kwargs.pop('''torchscript''' , self.torchscript ) __lowercase : str = kwargs.pop('''torch_xla_tpu_print_metrics''' , self.torch_xla_tpu_print_metrics ) __lowercase : str = kwargs.pop('''fp16_opt_level''' , self.fpaa_opt_level ) super().__init__(**UpperCamelCase_ ) UpperCamelCase =field(default=snake_case , metadata={"help": "Trace the models using torchscript"} ) UpperCamelCase =field(default=snake_case , metadata={"help": "Print Xla/PyTorch tpu metrics"} ) UpperCamelCase =field( default="O1" , metadata={ "help": ( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. " "See details at https://nvidia.github.io/apex/amp.html" ) } , ) @cached_property def _lowerCamelCase ( self ) -> Tuple["torch.device", int]: requires_backends(self , ['''torch'''] ) logger.info('''PyTorch: setting up devices''' ) if not self.cuda: __lowercase : str = torch.device('''cpu''' ) __lowercase : Optional[Any] = 0 elif is_torch_tpu_available(): __lowercase : str = xm.xla_device() __lowercase : Any = 0 else: __lowercase : List[str] = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowercase : Any = torch.cuda.device_count() return device, n_gpu @property def _lowerCamelCase ( self ) -> Union[str, Any]: return is_torch_tpu_available() and self.tpu @property def _lowerCamelCase ( self ) -> int: requires_backends(self , ['''torch'''] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def _lowerCamelCase ( self ) -> "torch.device": requires_backends(self , ['''torch'''] ) return self._setup_devices[0] @property def _lowerCamelCase ( self ) -> int: requires_backends(self , ['''torch'''] ) return self._setup_devices[1] @property def _lowerCamelCase ( self ) -> Dict: return self.n_gpu > 0
249
0
'''simple docstring''' import qiskit def __UpperCAmelCase ( a_: List[str], a_: Tuple ): _UpperCAmelCase : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register _UpperCAmelCase : List[str] = qiskit.QuantumCircuit(lowercase_, lowercase_ ) # Map the quantum measurement to the classical bits circuit.measure([0], [0] ) # Execute the circuit on the simulator _UpperCAmelCase : List[str] = qiskit.execute(lowercase_, lowercase_, shots=1_000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(lowercase_ ) if __name__ == "__main__": print(f'Total count for various states are: {single_qubit_measure(1, 1)}')
361
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __a = datasets.utils.logging.get_logger(__name__) __a = ['names', 'prefix'] __a = ['warn_bad_lines', 'error_bad_lines', 'mangle_dupe_cols'] __a = ['encoding_errors', 'on_bad_lines'] __a = ['date_format'] @dataclass class A__ ( datasets.BuilderConfig ): """simple docstring""" UpperCamelCase_ : str = "," UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : Optional[Union[int, List[int], str]] = "infer" UpperCamelCase_ : Optional[List[str]] = None UpperCamelCase_ : Optional[List[str]] = None UpperCamelCase_ : Optional[Union[int, str, List[int], List[str]]] = None UpperCamelCase_ : Optional[Union[List[int], List[str]]] = None UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : bool = True UpperCamelCase_ : Optional[Literal["c", "python", "pyarrow"]] = None UpperCamelCase_ : Dict[Union[int, str], Callable[[Any], Any]] = None UpperCamelCase_ : Optional[list] = None UpperCamelCase_ : Optional[list] = None UpperCamelCase_ : bool = False UpperCamelCase_ : Optional[Union[int, List[int]]] = None UpperCamelCase_ : Optional[int] = None UpperCamelCase_ : Optional[Union[str, List[str]]] = None UpperCamelCase_ : bool = True UpperCamelCase_ : bool = True UpperCamelCase_ : bool = False UpperCamelCase_ : bool = True UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : str = "." UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : str = '"' UpperCamelCase_ : int = 0 UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : bool = True UpperCamelCase_ : bool = True UpperCamelCase_ : int = 0 UpperCamelCase_ : bool = True UpperCamelCase_ : bool = False UpperCamelCase_ : Optional[str] = None UpperCamelCase_ : int = 1_00_00 UpperCamelCase_ : Optional[datasets.Features] = None UpperCamelCase_ : Optional[str] = "strict" UpperCamelCase_ : Literal["error", "warn", "skip"] = "error" UpperCamelCase_ : Optional[str] = None def _lowerCAmelCase ( self : str ) -> Tuple: """simple docstring""" if self.delimiter is not None: _UpperCAmelCase : Any = self.delimiter if self.column_names is not None: _UpperCAmelCase : List[Any] = self.column_names @property def _lowerCAmelCase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" _UpperCAmelCase : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A__ ( datasets.ArrowBasedBuilder ): """simple docstring""" UpperCamelCase_ : int = CsvConfig def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def _lowerCAmelCase ( self : Tuple , lowerCAmelCase__ : str ) -> List[str]: """simple docstring""" if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) _UpperCAmelCase : List[str] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): _UpperCAmelCase : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : Any = [files] _UpperCAmelCase : List[Any] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] _UpperCAmelCase : Optional[Any] = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : str = [files] _UpperCAmelCase : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _lowerCAmelCase ( self : List[Any] , lowerCAmelCase__ : pa.Table ) -> pa.Table: """simple docstring""" if self.config.features is not None: _UpperCAmelCase : Tuple = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast _UpperCAmelCase : Any = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example _UpperCAmelCase : int = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _lowerCAmelCase ( self : Dict , lowerCAmelCase__ : Dict ) -> Dict: """simple docstring""" _UpperCAmelCase : int = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str _UpperCAmelCase : Optional[Any] = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): _UpperCAmelCase : Optional[Any] = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): _UpperCAmelCase : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
17
0
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class lowercase_ ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self : Any ): _A = tempfile.mkdtemp() # fmt: off _A = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest'] # fmt: on _A = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) _A = { 'do_resize': True, 'size': {'height': 18, 'width': 18}, 'do_normalize': True, 'image_mean': [0.5, 0.5, 0.5], 'image_std': [0.5, 0.5, 0.5], } _A = os.path.join(self.tmpdirname , _UpperCAmelCase ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] , **_UpperCAmelCase : List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] , **_UpperCAmelCase : Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): shutil.rmtree(self.tmpdirname ) def lowerCAmelCase_ ( self : List[str] ): _A = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _A = [Image.fromarray(np.moveaxis(_UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCAmelCase_ ( self : Optional[Any] ): _A = self.get_tokenizer() _A = self.get_image_processor() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) _A = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] ): _A = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _A = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _A = self.get_image_processor(do_normalize=_UpperCAmelCase , padding_value=1.0 ) _A = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): _A = self.get_image_processor() _A = self.get_tokenizer() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) _A = self.prepare_image_inputs() _A = image_processor(_UpperCAmelCase , return_tensors='np' ) _A = processor(images=_UpperCAmelCase , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowerCAmelCase_ ( self : str ): _A = self.get_image_processor() _A = self.get_tokenizer() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) _A = 'lower newer' _A = processor(text=_UpperCAmelCase ) _A = tokenizer(_UpperCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCAmelCase_ ( self : Tuple ): _A = self.get_image_processor() _A = self.get_tokenizer() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) _A = 'lower newer' _A = self.prepare_image_inputs() _A = processor(text=_UpperCAmelCase , images=_UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'token_type_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with self.assertRaises(_UpperCAmelCase ): processor() def lowerCAmelCase_ ( self : List[Any] ): _A = self.get_image_processor() _A = self.get_tokenizer() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) _A = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _A = processor.batch_decode(_UpperCAmelCase ) _A = tokenizer.batch_decode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): _A = self.get_image_processor() _A = self.get_tokenizer() _A = VisionTextDualEncoderProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase ) _A = 'lower newer' _A = self.prepare_image_inputs() _A = processor(text=_UpperCAmelCase , images=_UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
315
"""simple docstring""" def _snake_case ( _snake_case : list , _snake_case : int = 0 ) -> list: '''simple docstring''' _A = length or len(_snake_case ) _A = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: _A , _A = list_data[i + 1], list_data[i] _A = True return list_data if not swapped else bubble_sort(_snake_case , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
315
1
"""simple docstring""" import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) a_ : Dict = [ 'cross_validation.py', 'gradient_accumulation.py', 'local_sgd.py', 'multi_process_metrics.py', 'memory.py', 'automatic_gradient_accumulation.py', 'fsdp_with_peak_mem_tracking.py', 'deepspeed_with_config_support.py', 'megatron_lm_gpt_pretraining.py', ] class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self , a , a , a = None , a = None) -> Optional[int]: SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = os.path.abspath(os.path.join('examples' , 'by_feature')) SCREAMING_SNAKE_CASE = os.path.abspath('examples') for item in os.listdir(a): if item not in EXCLUDE_EXAMPLES: SCREAMING_SNAKE_CASE = os.path.join(a , a) if os.path.isfile(a) and ".py" in item_path: with self.subTest( tested_script=a , feature_script=a , tested_section='main()' if parser_only else 'training_function()' , ): SCREAMING_SNAKE_CASE = compare_against_test( os.path.join(a , a) , a , a , a) SCREAMING_SNAKE_CASE = '\n'.join(a) if special_strings is not None: for string in special_strings: SCREAMING_SNAKE_CASE = diff.replace(a , '') self.assertEqual(a , '') def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: self.one_complete_example('complete_nlp_example.py' , a) self.one_complete_example('complete_nlp_example.py' , a) def SCREAMING_SNAKE_CASE__ ( self) -> Dict: SCREAMING_SNAKE_CASE = os.path.abspath(os.path.join('examples' , 'cv_example.py')) SCREAMING_SNAKE_CASE = [ ' ' * 16 + '{\n\n', ' ' * 20 + '"accuracy": eval_metric["accuracy"],\n\n', ' ' * 20 + '"f1": eval_metric["f1"],\n\n', ' ' * 20 + '"train_loss": total_loss.item() / len(train_dataloader),\n\n', ' ' * 20 + '"epoch": epoch,\n\n', ' ' * 16 + '},\n\n', ' ' * 16 + 'step=epoch,\n', ' ' * 12, ' ' * 8 + 'for step, batch in enumerate(active_dataloader):\n', ] self.one_complete_example('complete_cv_example.py' , a , a , a) self.one_complete_example('complete_cv_example.py' , a , a , a) @mock.patch.dict(os.environ , {'''TESTING_MOCKED_DATALOADERS''': '''1'''} ) class _snake_case ( A__ ): _lowercase : int = False @classmethod def SCREAMING_SNAKE_CASE__ ( cls) -> Union[str, Any]: super().setUpClass() SCREAMING_SNAKE_CASE = tempfile.mkdtemp() SCREAMING_SNAKE_CASE = os.path.join(cls._tmpdir , 'default_config.yml') write_basic_config(save_location=cls.configPath) SCREAMING_SNAKE_CASE = ['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def SCREAMING_SNAKE_CASE__ ( cls) -> Dict: super().tearDownClass() shutil.rmtree(cls._tmpdir) def SCREAMING_SNAKE_CASE__ ( self) -> Optional[Any]: SCREAMING_SNAKE_CASE = f''' examples/by_feature/checkpointing.py --checkpointing_steps epoch --output_dir {self.tmpdir} '''.split() run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'epoch_0'))) def SCREAMING_SNAKE_CASE__ ( self) -> Dict: SCREAMING_SNAKE_CASE = f''' examples/by_feature/checkpointing.py --checkpointing_steps 1 --output_dir {self.tmpdir} '''.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'step_2'))) def SCREAMING_SNAKE_CASE__ ( self) -> Any: SCREAMING_SNAKE_CASE = f''' examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'epoch_0')} '''.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=a) self.assertNotIn('epoch 0:' , a) self.assertIn('epoch 1:' , a) def SCREAMING_SNAKE_CASE__ ( self) -> int: SCREAMING_SNAKE_CASE = f''' examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'step_2')} '''.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=a) if torch.cuda.is_available(): SCREAMING_SNAKE_CASE = torch.cuda.device_count() else: SCREAMING_SNAKE_CASE = 1 if num_processes > 1: self.assertNotIn('epoch 0:' , a) self.assertIn('epoch 1:' , a) else: self.assertIn('epoch 0:' , a) self.assertIn('epoch 1:' , a) @slow def SCREAMING_SNAKE_CASE__ ( self) -> Any: SCREAMING_SNAKE_CASE = '\n examples/by_feature/cross_validation.py\n --num_folds 2\n '.split() with mock.patch.dict(os.environ , {'TESTING_MOCKED_DATALOADERS': '0'}): SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=a) SCREAMING_SNAKE_CASE = re.findall('({.+})' , a) SCREAMING_SNAKE_CASE = [r for r in results if 'accuracy' in r][-1] SCREAMING_SNAKE_CASE = ast.literal_eval(a) self.assertGreaterEqual(results['accuracy'] , 0.75) def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: SCREAMING_SNAKE_CASE = ['examples/by_feature/multi_process_metrics.py'] run_command(self._launch_args + testargs) @require_trackers @mock.patch.dict(os.environ , {'WANDB_MODE': 'offline'}) def SCREAMING_SNAKE_CASE__ ( self) -> Any: with tempfile.TemporaryDirectory() as tmpdir: SCREAMING_SNAKE_CASE = f''' examples/by_feature/tracking.py --with_tracking --project_dir {tmpdir} '''.split() run_command(self._launch_args + testargs) self.assertTrue(os.path.exists(os.path.join(a , 'tracking'))) def SCREAMING_SNAKE_CASE__ ( self) -> int: SCREAMING_SNAKE_CASE = ['examples/by_feature/gradient_accumulation.py'] run_command(self._launch_args + testargs) def SCREAMING_SNAKE_CASE__ ( self) -> Union[str, Any]: SCREAMING_SNAKE_CASE = ['examples/by_feature/local_sgd.py'] run_command(self._launch_args + testargs)
360
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging a_ : Union[str, Any] = logging.get_logger(__name__) class _snake_case ( A__ ): _lowercase : List[str] = ['''pixel_values'''] def __init__( self , a = True , a = 1 / 255 , a = True , a = 8 , **a , ) -> None: super().__init__(**a) SCREAMING_SNAKE_CASE = do_rescale SCREAMING_SNAKE_CASE = rescale_factor SCREAMING_SNAKE_CASE = do_pad SCREAMING_SNAKE_CASE = pad_size def SCREAMING_SNAKE_CASE__ ( self , a , a , a = None , **a) -> np.ndarray: return rescale(a , scale=a , data_format=a , **a) def SCREAMING_SNAKE_CASE__ ( self , a , a , a = None) -> List[str]: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = get_image_size(a) SCREAMING_SNAKE_CASE = (old_height // size + 1) * size - old_height SCREAMING_SNAKE_CASE = (old_width // size + 1) * size - old_width return pad(a , ((0, pad_height), (0, pad_width)) , mode='symmetric' , data_format=a) def SCREAMING_SNAKE_CASE__ ( self , a , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ) -> List[str]: SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE = do_pad if do_pad is not None else self.do_pad SCREAMING_SNAKE_CASE = pad_size if pad_size is not None else self.pad_size SCREAMING_SNAKE_CASE = make_list_of_images(a) if not valid_images(a): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.') # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(a) for image in images] if do_rescale: SCREAMING_SNAKE_CASE = [self.rescale(image=a , scale=a) for image in images] if do_pad: SCREAMING_SNAKE_CASE = [self.pad(a , size=a) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(a , a) for image in images] SCREAMING_SNAKE_CASE = {'pixel_values': images} return BatchFeature(data=a , tensor_type=a)
327
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class _A ( metaclass=lowerCAmelCase ): snake_case__ : Optional[int] = ['torch', 'torchsde'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """torchsde"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """torchsde"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """torchsde"""] )
197
"""simple docstring""" from __future__ import annotations from scipy.special import comb # type: ignore class _A : def __init__( self , __lowerCAmelCase ): """simple docstring""" lowercase = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. lowercase = len(__lowerCAmelCase ) - 1 def A__ ( self , __lowerCAmelCase ): """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." lowercase = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __lowerCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__lowerCAmelCase ) , 5 ) == 1 return output_values def A__ ( self , __lowerCAmelCase ): """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." lowercase = self.basis_function(__lowerCAmelCase ) lowercase = 0.0 lowercase = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def A__ ( self , __lowerCAmelCase = 0.0_1 ): """simple docstring""" from matplotlib import pyplot as plt # type: ignore lowercase = [] # x coordinates of points to plot lowercase = [] # y coordinates of points to plot lowercase = 0.0 while t <= 1: lowercase = self.bezier_curve_function(__lowerCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size lowercase = [i[0] for i in self.list_of_points] lowercase = [i[1] for i in self.list_of_points] plt.plot( __lowerCAmelCase , __lowerCAmelCase , color="""blue""" , label="""Curve of Degree """ + str(self.degree ) , ) plt.scatter(__lowerCAmelCase , __lowerCAmelCase , color="""red""" , label="""Control Points""" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
197
1
'''simple docstring''' __snake_case = 65521 def a ( __a ) -> int: '''simple docstring''' UpperCamelCase__ :Tuple = 1 UpperCamelCase__ :Tuple = 0 for plain_chr in plain_text: UpperCamelCase__ :int = (a + ord(__a )) % MOD_ADLER UpperCamelCase__ :Tuple = (b + a) % MOD_ADLER return (b << 16) | a
219
'''simple docstring''' from __future__ import annotations from typing import Any def a ( __a ) -> None: '''simple docstring''' create_state_space_tree(__a , [] , 0 ) def a ( __a , __a , __a ) -> None: '''simple docstring''' if index == len(__a ): print(__a ) return create_state_space_tree(__a , __a , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__a , __a , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __snake_case = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
219
1
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' assert ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and number_of_steps > 0 ), f"""number_of_steps needs to be positive integer, your input {number_of_steps}""" if number_of_steps == 1: return 1 __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 1, 1 for _ in range(number_of_steps - 1 ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
54
"""simple docstring""" import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( "kwargs, expected" , [ ({"num_shards": 0, "max_num_jobs": 1}, []), ({"num_shards": 10, "max_num_jobs": 1}, [range(10 )]), ({"num_shards": 10, "max_num_jobs": 10}, [range(lowerCAmelCase_ , i + 1 ) for i in range(10 )]), ({"num_shards": 1, "max_num_jobs": 10}, [range(1 )]), ({"num_shards": 10, "max_num_jobs": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({"num_shards": 3, "max_num_jobs": 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, max_num_jobs, expected" , [ ({"foo": 0}, 10, [{"foo": 0}]), ({"shards": [0, 1, 2, 3]}, 1, [{"shards": [0, 1, 2, 3]}]), ({"shards": [0, 1, 2, 3]}, 4, [{"shards": [0]}, {"shards": [1]}, {"shards": [2]}, {"shards": [3]}]), ({"shards": [0, 1]}, 4, [{"shards": [0]}, {"shards": [1]}]), ({"shards": [0, 1, 2, 3]}, 2, [{"shards": [0, 1]}, {"shards": [2, 3]}]), ] , ) def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = _split_gen_kwargs(lowerCAmelCase_ , lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, expected" , [ ({"foo": 0}, 1), ({"shards": [0]}, 1), ({"shards": [0, 1, 2, 3]}, 4), ({"shards": [0, 1, 2, 3], "foo": 0}, 4), ({"shards": [0, 1, 2, 3], "other": (0, 1)}, 4), ({"shards": [0, 1, 2, 3], "shards2": [0, 1]}, RuntimeError), ] , ) def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: __SCREAMING_SNAKE_CASE = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
54
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Any: lowercase : Any = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __snake_case ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , unittest.TestCase ): _a : Any= StableDiffusionLatentUpscalePipeline _a : List[str]= TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "height", "width", "cross_attention_kwargs", "negative_prompt_embeds", "prompt_embeds", } _a : Optional[Any]= PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"} _a : Tuple= TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _a : int= frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _a : Optional[int]= frozenset([] ) _a : List[str]= True @property def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Dict = 1 lowercase : Dict = 4 lowercase : Union[str, Any] = (16, 16) lowercase : Optional[Any] = floats_tensor((batch_size, num_channels) + sizes ,rng=random.Random(0 ) ).to(snake_case ) return image def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' torch.manual_seed(0 ) lowercase : Union[str, Any] = UNetaDConditionModel( act_fn="""gelu""" ,attention_head_dim=8 ,norm_num_groups=snake_case ,block_out_channels=[32, 32, 64, 64] ,time_cond_proj_dim=160 ,conv_in_kernel=1 ,conv_out_kernel=1 ,cross_attention_dim=32 ,down_block_types=( """KDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", ) ,in_channels=8 ,mid_block_type=snake_case ,only_cross_attention=snake_case ,out_channels=5 ,resnet_time_scale_shift="""scale_shift""" ,time_embedding_type="""fourier""" ,timestep_post_act="""gelu""" ,up_block_types=("""KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KUpBlock2D""") ,) lowercase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 32, 64, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=[ """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", ] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,) lowercase : int = EulerDiscreteScheduler(prediction_type="""sample""" ) lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,hidden_act="""quick_gelu""" ,projection_dim=512 ,) lowercase : str = CLIPTextModel(snake_case ) lowercase : List[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowercase : Optional[int] = { """unet""": model.eval(), """vae""": vae.eval(), """scheduler""": scheduler, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case=0 ): '''simple docstring''' if str(snake_case ).startswith("""mps""" ): lowercase : Tuple = torch.manual_seed(snake_case ) else: lowercase : Tuple = torch.Generator(device=snake_case ).manual_seed(snake_case ) lowercase : Optional[int] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": self.dummy_image.cpu(), """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : List[str] = """cpu""" lowercase : List[Any] = self.get_dummy_components() lowercase : Dict = self.pipeline_class(**snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) lowercase : Optional[Any] = self.get_dummy_inputs(snake_case ) lowercase : Any = pipe(**snake_case ).images lowercase : Dict = image[0, -3:, -3:, -1] self.assertEqual(image.shape ,(1, 256, 256, 3) ) lowercase : Optional[int] = np.array( [0.47_222_412, 0.41_921_633, 0.44_717_434, 0.46_874_192, 0.42_588_258, 0.46_150_726, 0.4_677_534, 0.45_583_832, 0.48_579_055] ) lowercase : List[Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case ,1e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_save_load_local(expected_max_difference=3e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Tuple = [ """DDIMScheduler""", """DDPMScheduler""", """PNDMScheduler""", """HeunDiscreteScheduler""", """EulerAncestralDiscreteScheduler""", """KDPM2DiscreteScheduler""", """KDPM2AncestralDiscreteScheduler""", """DPMSolverSDEScheduler""", ] lowercase : List[Any] = self.get_dummy_components() lowercase : Tuple = self.pipeline_class(**snake_case ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) lowercase : Tuple = self.get_dummy_inputs(snake_case ) lowercase : str = 2 lowercase : Dict = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue lowercase : Tuple = getattr(snake_case ,scheduler_enum.name ) lowercase : int = scheduler_cls.from_config(pipe.scheduler.config ) lowercase : Dict = pipe(**snake_case )[0] outputs.append(snake_case ) assert check_same_shape(snake_case ) @require_torch_gpu @slow class __snake_case ( unittest.TestCase ): def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Optional[Any] = torch.manual_seed(33 ) lowercase : Union[str, Any] = StableDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" ,torch_dtype=torch.floataa ) pipe.to("""cuda""" ) lowercase : List[str] = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" ,torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) lowercase : List[str] = """a photo of an astronaut high resolution, unreal engine, ultra realistic""" lowercase : Any = pipe(snake_case ,generator=snake_case ,output_type="""latent""" ).images lowercase : Union[str, Any] = upscaler( prompt=snake_case ,image=snake_case ,num_inference_steps=20 ,guidance_scale=0 ,generator=snake_case ,output_type="""np""" ,).images[0] lowercase : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy""" ) assert np.abs((expected_image - image).mean() ) < 5e-2 def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : List[str] = torch.manual_seed(33 ) lowercase : Tuple = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" ,torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) lowercase : str = """the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas""" lowercase : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png""" ) lowercase : List[Any] = upscaler( prompt=snake_case ,image=snake_case ,num_inference_steps=20 ,guidance_scale=0 ,generator=snake_case ,output_type="""np""" ,).images[0] lowercase : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy""" ) assert np.abs((expected_image - image).max() ) < 5e-2
285
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( lowerCAmelCase , unittest.TestCase ): _a : Optional[int]= None _a : Optional[Any]= BloomTokenizerFast _a : Tuple= BloomTokenizerFast _a : str= True _a : Optional[int]= False _a : List[Any]= "tokenizer_file" _a : List[Any]= {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().setUp() lowercase : Optional[Any] = BloomTokenizerFast.from_pretrained("""bigscience/tokenizer""" ) tokenizer.save_pretrained(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self ,**snake_case ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname ,**snake_case ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Tuple = self.get_rust_tokenizer() lowercase : List[str] = ["""The quick brown fox</s>""", """jumps over the lazy dog</s>"""] lowercase : Optional[int] = [[2175, 23714, 73173, 144252, 2], [77, 132619, 3478, 368, 109586, 35433, 2]] lowercase : Any = tokenizer.batch_encode_plus(snake_case )["""input_ids"""] self.assertListEqual(snake_case ,snake_case ) lowercase : Optional[int] = tokenizer.batch_decode(snake_case ) self.assertListEqual(snake_case ,snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case=6 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): lowercase : Dict = self.rust_tokenizer_class.from_pretrained(snake_case ,**snake_case ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input lowercase : Dict = """This is a simple input""" lowercase : Tuple = ["""This is a simple input 1""", """This is a simple input 2"""] lowercase : Dict = ("""This is a simple input""", """This is a pair""") lowercase : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests try: tokenizer_r.encode(snake_case ,max_length=snake_case ) tokenizer_r.encode_plus(snake_case ,max_length=snake_case ) tokenizer_r.batch_encode_plus(snake_case ,max_length=snake_case ) tokenizer_r.encode(snake_case ,max_length=snake_case ) tokenizer_r.batch_encode_plus(snake_case ,max_length=snake_case ) except ValueError: self.fail("""Bloom Tokenizer should be able to deal with padding""" ) lowercase : Optional[int] = None # Hotfixing padding = None self.assertRaises(snake_case ,tokenizer_r.encode ,snake_case ,max_length=snake_case ,padding="""max_length""" ) # Simple input self.assertRaises(snake_case ,tokenizer_r.encode_plus ,snake_case ,max_length=snake_case ,padding="""max_length""" ) # Simple input self.assertRaises( snake_case ,tokenizer_r.batch_encode_plus ,snake_case ,max_length=snake_case ,padding="""max_length""" ,) # Pair input self.assertRaises(snake_case ,tokenizer_r.encode ,snake_case ,max_length=snake_case ,padding="""max_length""" ) # Pair input self.assertRaises(snake_case ,tokenizer_r.encode_plus ,snake_case ,max_length=snake_case ,padding="""max_length""" ) # Pair input self.assertRaises( snake_case ,tokenizer_r.batch_encode_plus ,snake_case ,max_length=snake_case ,padding="""max_length""" ,) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : List[Any] = self.get_rust_tokenizer() lowercase : List[str] = load_dataset("""xnli""" ,"""all_languages""" ,split="""test""" ,streaming=snake_case ) lowercase : Optional[Any] = next(iter(snake_case ) )["""premise"""] # pick up one data lowercase : str = list(sample_data.values() ) lowercase : Optional[int] = list(map(tokenizer.encode ,snake_case ) ) lowercase : Dict = [tokenizer.decode(snake_case ,clean_up_tokenization_spaces=snake_case ) for x in output_tokens] self.assertListEqual(snake_case ,snake_case ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) ,1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) ,1 )
285
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ = { '''configuration_git''': ['''GIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GitConfig''', '''GitVisionConfig'''], '''processing_git''': ['''GitProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''GIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GitForCausalLM''', '''GitModel''', '''GitPreTrainedModel''', '''GitVisionModel''', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
8
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __SCREAMING_SNAKE_CASE (*SCREAMING_SNAKE_CASE__ ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): snake_case_ = list(SCREAMING_SNAKE_CASE__ ) for i in range(len(SCREAMING_SNAKE_CASE__ ) ): snake_case_ = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): snake_case_ = [ '''CUDA out of memory.''', # CUDA OOM '''cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.''', # CUDNN SNAFU '''DefaultCPUAllocator: can\'t allocate memory''', # CPU OOM ] if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = 128 ): if function is None: return functools.partial(SCREAMING_SNAKE_CASE__ , starting_batch_size=SCREAMING_SNAKE_CASE__ ) snake_case_ = starting_batch_size def decorator(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() snake_case_ = list(inspect.signature(SCREAMING_SNAKE_CASE__ ).parameters.keys() ) # Guard against user error if len(SCREAMING_SNAKE_CASE__ ) < (len(SCREAMING_SNAKE_CASE__ ) + 1): snake_case_ = ''', '''.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError('''No executable batch size found, reached zero.''' ) try: return function(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) except Exception as e: if should_reduce_batch_size(SCREAMING_SNAKE_CASE__ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
8
1
import cmath import math def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> complex: """simple docstring""" _snake_case = math.radians(_UpperCamelCase ) _snake_case = math.radians(_UpperCamelCase ) # Convert voltage and current to rectangular form _snake_case = cmath.rect(_UpperCamelCase , _UpperCamelCase ) _snake_case = cmath.rect(_UpperCamelCase , _UpperCamelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
370
__A = { '''Pillow''': '''Pillow<10.0.0''', '''accelerate''': '''accelerate>=0.20.3''', '''av''': '''av==9.2.0''', '''beautifulsoup4''': '''beautifulsoup4''', '''black''': '''black~=23.1''', '''codecarbon''': '''codecarbon==1.2.0''', '''cookiecutter''': '''cookiecutter==1.7.3''', '''dataclasses''': '''dataclasses''', '''datasets''': '''datasets!=2.5.0''', '''decord''': '''decord==0.6.0''', '''deepspeed''': '''deepspeed>=0.9.3''', '''diffusers''': '''diffusers''', '''dill''': '''dill<0.3.5''', '''evaluate''': '''evaluate>=0.2.0''', '''fairscale''': '''fairscale>0.3''', '''faiss-cpu''': '''faiss-cpu''', '''fastapi''': '''fastapi''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1,<=0.7.0''', '''ftfy''': '''ftfy''', '''fugashi''': '''fugashi>=1.0''', '''GitPython''': '''GitPython<3.1.19''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.14.1,<1.0''', '''importlib_metadata''': '''importlib_metadata''', '''ipadic''': '''ipadic>=1.0.0,<2.0''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2,<=0.4.13''', '''jaxlib''': '''jaxlib>=0.1.65,<=0.4.13''', '''jieba''': '''jieba''', '''kenlm''': '''kenlm''', '''keras-nlp''': '''keras-nlp>=0.3.1''', '''librosa''': '''librosa''', '''nltk''': '''nltk''', '''natten''': '''natten>=0.14.6''', '''numpy''': '''numpy>=1.17''', '''onnxconverter-common''': '''onnxconverter-common''', '''onnxruntime-tools''': '''onnxruntime-tools>=1.4.2''', '''onnxruntime''': '''onnxruntime>=1.4.0''', '''opencv-python''': '''opencv-python''', '''optuna''': '''optuna''', '''optax''': '''optax>=0.0.8,<=0.1.4''', '''packaging''': '''packaging>=20.0''', '''parameterized''': '''parameterized''', '''phonemizer''': '''phonemizer''', '''protobuf''': '''protobuf''', '''psutil''': '''psutil''', '''pyyaml''': '''pyyaml>=5.1''', '''pydantic''': '''pydantic<2''', '''pytest''': '''pytest>=7.2.0''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''python''': '''python>=3.8.0''', '''ray[tune]''': '''ray[tune]''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''rhoknp''': '''rhoknp>=1.1.0,<1.3.1''', '''rjieba''': '''rjieba''', '''rouge-score''': '''rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1''', '''ruff''': '''ruff>=0.0.241,<=0.0.259''', '''sacrebleu''': '''sacrebleu>=1.4.12,<2.0.0''', '''sacremoses''': '''sacremoses''', '''safetensors''': '''safetensors>=0.3.1''', '''sagemaker''': '''sagemaker>=2.31.0''', '''scikit-learn''': '''scikit-learn''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''sigopt''': '''sigopt''', '''starlette''': '''starlette''', '''sudachipy''': '''sudachipy>=0.6.6''', '''sudachidict_core''': '''sudachidict_core>=20220729''', '''tensorflow-cpu''': '''tensorflow-cpu>=2.6,<2.14''', '''tensorflow''': '''tensorflow>=2.6,<2.14''', '''tensorflow-text''': '''tensorflow-text<2.14''', '''tf2onnx''': '''tf2onnx''', '''timeout-decorator''': '''timeout-decorator''', '''timm''': '''timm''', '''tokenizers''': '''tokenizers>=0.11.1,!=0.11.3,<0.14''', '''torch''': '''torch>=1.9,!=1.12.0''', '''torchaudio''': '''torchaudio''', '''torchvision''': '''torchvision''', '''pyctcdecode''': '''pyctcdecode>=0.4.0''', '''tqdm''': '''tqdm>=4.27''', '''unidic''': '''unidic>=1.0.2''', '''unidic_lite''': '''unidic_lite>=1.0.7''', '''urllib3''': '''urllib3<2.0.0''', '''uvicorn''': '''uvicorn''', }
278
0
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _lowerCAmelCase : @staticmethod def __a ( *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: pass @is_pipeline_test @require_vision class _lowerCAmelCase ( unittest.TestCase ): @require_torch def __a ( self ) -> Dict: lowerCAmelCase_ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , ) lowerCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCAmelCase_ = image_classifier(lowerCAmelCase__ , candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(lowerCAmelCase__ ) , [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ] , ) lowerCAmelCase_ = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], ] , ) @require_tf def __a ( self ) -> int: lowerCAmelCase_ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , framework="tf" ) lowerCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCAmelCase_ = image_classifier(lowerCAmelCase__ , candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}] , ) lowerCAmelCase_ = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], [ {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, {"score": 0.333, "label": ANY(lowerCAmelCase__ )}, ], ] , ) @slow @require_torch def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCAmelCase_ = image_classifier(lowerCAmelCase__ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) lowerCAmelCase_ = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , ) @slow @require_tf def __a ( self ) -> Optional[Any]: lowerCAmelCase_ = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , framework="tf" ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCAmelCase_ = image_classifier(lowerCAmelCase__ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) lowerCAmelCase_ = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , )
231
"""simple docstring""" import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def _SCREAMING_SNAKE_CASE ( _lowercase : int = 8 ) ->str: '''simple docstring''' a : List[Any] = ascii_letters + digits + punctuation return "".join(secrets.choice(_lowercase ) for _ in range(_lowercase ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int ) ->str: '''simple docstring''' i -= len(_lowercase ) a : List[str] = i // 3 a : Any = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) a : int = ( chars_incl + random(_lowercase , quotient + remainder ) + random(_lowercase , _lowercase ) + random(_lowercase , _lowercase ) ) a : List[str] = list(_lowercase ) shuffle(_lowercase ) return "".join(_lowercase ) # random is a generalised function for letters, characters and numbers def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int ) ->str: '''simple docstring''' return "".join(secrets.choice(_lowercase ) for _ in range(_lowercase ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : int ) ->List[str]: '''simple docstring''' pass # Put your code here... def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Optional[int] ) ->int: '''simple docstring''' pass # Put your code here... def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : Optional[Any] ) ->Any: '''simple docstring''' pass # Put your code here... def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int = 8 ) ->bool: '''simple docstring''' if len(_lowercase ) < min_length: # Your Password must be at least 8 characters long return False a : List[str] = any(char in ascii_uppercase for char in password ) a : Optional[int] = any(char in ascii_lowercase for char in password ) a : List[str] = any(char in digits for char in password ) a : int = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def _SCREAMING_SNAKE_CASE ( ) ->Union[str, Any]: '''simple docstring''' a : Dict = int(input("Please indicate the max length of your password: " ).strip() ) a : str = input( "Please indicate the characters that must be in your password: " ).strip() print("Password generated:" , password_generator(_lowercase ) ) print( "Alternative Password generated:" , alternative_password_generator(_lowercase , _lowercase ) , ) print("[If you are thinking of using this passsword, You better save it.]" ) if __name__ == "__main__": main()
105
0
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _UpperCAmelCase (UpperCamelCase_ : str , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any] ): '''simple docstring''' # Initialise PyTorch model _lowerCAmelCase : Any = TaConfig.from_json_file(UpperCamelCase_ ) print(F"Building PyTorch model from configuration: {config}" ) _lowerCAmelCase : List[Any] = TaForConditionalGeneration(UpperCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase_ ) if __name__ == "__main__": _lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _lowerCamelCase : int = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
159
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[int] = ["ConditionalDetrFeatureExtractor"] _lowerCamelCase : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys _lowerCamelCase : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
159
1
"""simple docstring""" from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets a_ = """\ @inproceedings{wang2019glue, title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding}, author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.}, note={In the Proceedings of ICLR.}, year={2019} } """ a_ = """\ GLUE, the General Language Understanding Evaluation benchmark (https://gluebenchmark.com/) is a collection of resources for training, evaluating, and analyzing natural language understanding systems. """ a_ = """ Compute GLUE evaluation metric associated to each GLUE dataset. Args: predictions: list of predictions to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. Returns: depending on the GLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"pearson\": Pearson Correlation \"spearmanr\": Spearman Correlation \"matthews_correlation\": Matthew Correlation Examples: >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp' >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> glue_metric = datasets.load_metric('glue', 'stsb') >>> references = [0., 1., 2., 3., 4., 5.] >>> predictions = [0., 1., 2., 3., 4., 5.] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)}) {'pearson': 1.0, 'spearmanr': 1.0} >>> glue_metric = datasets.load_metric('glue', 'cola') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def __lowercase ( snake_case_ : Union[str, Any] ,snake_case_ : List[str] ) ->str: '''simple docstring''' return float((preds == labels).mean() ) def __lowercase ( snake_case_ : Any ,snake_case_ : List[str] ) ->List[str]: '''simple docstring''' __A : Union[str, Any] = simple_accuracy(snake_case_ ,snake_case_ ) __A : Optional[int] = float(fa_score(y_true=snake_case_ ,y_pred=snake_case_ ) ) return { "accuracy": acc, "f1": fa, } def __lowercase ( snake_case_ : List[str] ,snake_case_ : str ) ->Optional[Any]: '''simple docstring''' __A : int = float(pearsonr(snake_case_ ,snake_case_ )[0] ) __A : Dict = float(spearmanr(snake_case_ ,snake_case_ )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): """simple docstring""" def UpperCamelCase__( self ): '''simple docstring''' if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["sst2", "mnli", "mnli_mismatched", "mnli_matched", ''' '''"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' if self.config_name != '''stsb''' else '''float32''' ), '''references''': datasets.Value('''int64''' if self.config_name != '''stsb''' else '''float32''' ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' , ) def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "stsb": return pearson_and_spearman(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["sst2", "mnli", "mnli_mismatched", "mnli_matched", ''' '''"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]''' )
179
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch a_ = logging.get_logger(__name__) class __snake_case : """simple docstring""" def __init__( self , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase=None , __lowerCamelCase=None ): '''simple docstring''' if not conversation_id: __A : List[Any] = uuid.uuida() if past_user_inputs is None: __A : List[str] = [] if generated_responses is None: __A : Tuple = [] __A : uuid.UUID = conversation_id __A : List[str] = past_user_inputs __A : List[str] = generated_responses __A : Optional[str] = text def __eq__( self , __lowerCamelCase ): '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase = False ): '''simple docstring''' if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) __A : str = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: __A : Union[str, Any] = text def UpperCamelCase__( self ): '''simple docstring''' if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __A : List[Any] = None def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' self.generated_responses.append(__lowerCamelCase ) def UpperCamelCase__( self ): '''simple docstring''' for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ): '''simple docstring''' __A : Optional[Any] = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): __A : Tuple = '''user''' if is_user else '''bot''' output += F"""{name} >> {text} \n""" return output @add_end_docstrings( SCREAMING_SNAKE_CASE__ , R""" min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. """ , ) class __snake_case ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , *__lowerCamelCase , **__lowerCamelCase ): '''simple docstring''' super().__init__(*__lowerCamelCase , **__lowerCamelCase ) if self.tokenizer.pad_token_id is None: __A : Union[str, Any] = self.tokenizer.eos_token def UpperCamelCase__( self , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase ): '''simple docstring''' __A : str = {} __A : List[str] = {} __A : Any = {} if min_length_for_response is not None: __A : int = min_length_for_response if minimum_tokens is not None: __A : Any = minimum_tokens if "max_length" in generate_kwargs: __A : List[Any] = generate_kwargs['''max_length'''] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __A : str = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(__lowerCamelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self , __lowerCamelCase , __lowerCamelCase=0 , **__lowerCamelCase ): '''simple docstring''' __A : Any = super().__call__(__lowerCamelCase , num_workers=__lowerCamelCase , **__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) == 1: return outputs[0] return outputs def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase=32 ): '''simple docstring''' if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise ValueError('''ConversationalPipeline, expects Conversation as inputs''' ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ '''Add user inputs with the conversation\'s `add_user_input` method''' ) if hasattr(self.tokenizer , '''_build_conversation_input_ids''' ): __A : List[Any] = self.tokenizer._build_conversation_input_ids(__lowerCamelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version __A : int = self._legacy_parse_and_tokenize(__lowerCamelCase ) if self.framework == "pt": __A : Union[str, Any] = torch.LongTensor([input_ids] ) elif self.framework == "tf": __A : int = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase=10 , **__lowerCamelCase ): '''simple docstring''' __A : Tuple = generate_kwargs.get('''max_length''' , self.model.config.max_length ) __A : str = model_inputs['''input_ids'''].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) __A : str = max_length - minimum_tokens __A : Any = model_inputs['''input_ids'''][:, -trim:] if "attention_mask" in model_inputs: __A : Union[str, Any] = model_inputs['''attention_mask'''][:, -trim:] __A : Dict = model_inputs.pop('''conversation''' ) __A : List[str] = max_length __A : Dict = self.model.generate(**__lowerCamelCase , **__lowerCamelCase ) if self.model.config.is_encoder_decoder: __A : Any = 1 else: __A : List[Any] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase=True ): '''simple docstring''' __A : int = model_outputs['''output_ids'''] __A : Optional[int] = self.tokenizer.decode( output_ids[0] , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase , ) __A : Dict = model_outputs['''conversation'''] conversation.mark_processed() conversation.append_response(__lowerCamelCase ) return conversation def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : Tuple = self.tokenizer.eos_token_id __A : List[str] = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) ) if len(__lowerCamelCase ) > self.tokenizer.model_max_length: __A : List[str] = input_ids[-self.tokenizer.model_max_length :] return input_ids
179
1
'''simple docstring''' import math def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = 0 __lowercase = 0 while num > 0: __lowercase = num % 8 __lowercase = octal + (remainder * math.floor(math.pow(1_0 , lowerCamelCase_ ) )) counter += 1 __lowercase = math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return f"0o{int(lowerCamelCase_ )}" def _lowerCAmelCase ( ): print('''\n2 in octal is:''' ) print(decimal_to_octal(2 ) ) # = 2 print('''\n8 in octal is:''' ) print(decimal_to_octal(8 ) ) # = 10 print('''\n65 in octal is:''' ) print(decimal_to_octal(6_5 ) ) # = 101 print('''\n216 in octal is:''' ) print(decimal_to_octal(2_1_6 ) ) # = 330 print('''\n512 in octal is:''' ) print(decimal_to_octal(5_1_2 ) ) # = 1000 print('''\n''' ) if __name__ == "__main__": main()
217
'''simple docstring''' import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging _SCREAMING_SNAKE_CASE = ['''bart.large''', '''bart.large.mnli''', '''bart.large.cnn''', '''bart_xsum/model.pt'''] _SCREAMING_SNAKE_CASE = {'''bart.large''': BartModel, '''bart.large.mnli''': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('''0.9.0'''): raise Exception('''requires fairseq >= 0.9.0''') logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = ''' Hello world! cécé herlolip''' _SCREAMING_SNAKE_CASE = [ ('''model.classification_heads.mnli.dense.weight''', '''classification_head.dense.weight'''), ('''model.classification_heads.mnli.dense.bias''', '''classification_head.dense.bias'''), ('''model.classification_heads.mnli.out_proj.weight''', '''classification_head.out_proj.weight'''), ('''model.classification_heads.mnli.out_proj.bias''', '''classification_head.out_proj.bias'''), ] def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCamelCase_ , lowerCamelCase_ ) def _lowerCAmelCase ( lowerCamelCase_ : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : Union[str, Any] ): __lowercase = dct.pop(lowerCamelCase_ ) __lowercase = val def _lowerCAmelCase ( lowerCamelCase_ : Any ): __lowercase = torch.load(lowerCamelCase_ , map_location='''cpu''' ) __lowercase = torch.hub.load('''pytorch/fairseq''' , '''bart.large.cnn''' ).eval() hub_interface.model.load_state_dict(sd['''model'''] ) return hub_interface def _lowerCAmelCase ( lowerCamelCase_ : List[str] ): __lowercase , __lowercase = emb.weight.shape __lowercase = nn.Linear(lowerCamelCase_ , lowerCamelCase_ , bias=lowerCamelCase_ ) __lowercase = emb.weight.data return lin_layer @torch.no_grad() def _lowerCAmelCase ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : List[Any]=None ): if not os.path.exists(lowerCamelCase_ ): __lowercase = torch.hub.load('''pytorch/fairseq''' , lowerCamelCase_ ).eval() else: __lowercase = load_xsum_checkpoint(lowerCamelCase_ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowercase = checkpoint_path.replace('''.''' , '''-''' ) __lowercase = BartConfig.from_pretrained(lowerCamelCase_ ) __lowercase = bart.encode(lowerCamelCase_ ).unsqueeze(0 ) __lowercase = BartTokenizer.from_pretrained(lowerCamelCase_ ).encode(lowerCamelCase_ , return_tensors='''pt''' ).unsqueeze(0 ) if not torch.eq(lowerCamelCase_ , lowerCamelCase_ ).all(): raise ValueError( f"converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}" ) if checkpoint_path == "bart.large.mnli": __lowercase = bart.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''model.decoder.embed_tokens.weight'''] for src, dest in mnli_rename_keys: rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) __lowercase = BartForSequenceClassification(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = bart.predict('''mnli''' , lowerCamelCase_ , return_logits=lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ )[0] # logits else: # no classification heads to worry about __lowercase = bart.model.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''decoder.embed_tokens.weight'''] __lowercase = bart.extract_features(lowerCamelCase_ ) if hf_checkpoint_name == "facebook/bart-large": __lowercase = BartModel(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ ).model[0] else: __lowercase = BartForConditionalGeneration(lowerCamelCase_ ).eval() # an existing summarization ckpt model.model.load_state_dict(lowerCamelCase_ ) if hasattr(lowerCamelCase_ , '''lm_head''' ): __lowercase = make_linear_from_emb(model.model.shared ) __lowercase = model.model(lowerCamelCase_ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f"`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}" ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError('''Some values in `fairseq_output` are different from `new_model_outputs`''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default=None, type=str, help='''Which huggingface architecture to use: bart-large-xsum''' ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
217
1
import numpy as np def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ = 1e-12 , lowercase__ = 100 , ): assert np.shape(lowercase__ )[0] == np.shape(lowercase__ )[1] # Ensure proper dimensionality. assert np.shape(lowercase__ )[0] == np.shape(lowercase__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase__ ) == np.iscomplexobj(lowercase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = np.iscomplexobj(lowercase__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __SCREAMING_SNAKE_CASE : Dict = False __SCREAMING_SNAKE_CASE : Optional[int] = 0 __SCREAMING_SNAKE_CASE : Any = 0 __SCREAMING_SNAKE_CASE : Dict = 1e12 while not convergence: # Multiple matrix by the vector. __SCREAMING_SNAKE_CASE : Optional[int] = np.dot(lowercase__ , lowercase__ ) # Normalize the resulting output vector. __SCREAMING_SNAKE_CASE : int = w / np.linalg.norm(lowercase__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __SCREAMING_SNAKE_CASE : str = vector.conj().T if is_complex else vector.T __SCREAMING_SNAKE_CASE : Optional[Any] = np.dot(lowercase__ , np.dot(lowercase__ , lowercase__ ) ) # Check convergence. __SCREAMING_SNAKE_CASE : int = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __SCREAMING_SNAKE_CASE : Optional[Any] = True __SCREAMING_SNAKE_CASE : Tuple = lambda_ if is_complex: __SCREAMING_SNAKE_CASE : Tuple = np.real(lambda_ ) return lambda_, vector def _UpperCamelCase ( ): __SCREAMING_SNAKE_CASE : Union[str, Any] = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __SCREAMING_SNAKE_CASE : Tuple = np.array([41, 4, 20] ) __SCREAMING_SNAKE_CASE : Tuple = real_input_matrix.astype(np.complexaaa ) __SCREAMING_SNAKE_CASE : str = np.triu(1J * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __SCREAMING_SNAKE_CASE : Union[str, Any] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __SCREAMING_SNAKE_CASE : Any = real_input_matrix __SCREAMING_SNAKE_CASE : Optional[Any] = real_vector elif problem_type == "complex": __SCREAMING_SNAKE_CASE : List[str] = complex_input_matrix __SCREAMING_SNAKE_CASE : str = complex_vector # Our implementation. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : int = power_iteration(lowercase__ , lowercase__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[int] = np.linalg.eigh(lowercase__ ) # Last eigenvalue is the maximum one. __SCREAMING_SNAKE_CASE : List[str] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __SCREAMING_SNAKE_CASE : Optional[int] = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase__ ) - np.abs(lowercase__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
9
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) SCREAMING_SNAKE_CASE : List[str] = { "configuration_speech_to_text": ["SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Speech2TextConfig"], "processing_speech_to_text": ["Speech2TextProcessor"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["Speech2TextTokenizer"] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = ["Speech2TextFeatureExtractor"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Optional[Any] = [ "TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFSpeech2TextForConditionalGeneration", "TFSpeech2TextModel", "TFSpeech2TextPreTrainedModel", ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : str = [ "SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "Speech2TextForConditionalGeneration", "Speech2TextModel", "Speech2TextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
21
0
'''simple docstring''' import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __lowerCAmelCase : List[str] =logging.get_logger(__name__) __lowerCAmelCase : List[str] ={"vocab_file": "spiece.model"} __lowerCAmelCase : Optional[int] ={ "vocab_file": { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model", } } __lowerCAmelCase : List[Any] ={ "xlnet-base-cased": None, "xlnet-large-cased": None, } # Segments (not really needed) __lowerCAmelCase : List[str] =0 __lowerCAmelCase : int =1 __lowerCAmelCase : Optional[int] =2 __lowerCAmelCase : Optional[int] =3 __lowerCAmelCase : List[Any] =4 class UpperCAmelCase ( UpperCamelCase__ ): __lowercase = VOCAB_FILES_NAMES __lowercase = PRETRAINED_VOCAB_FILES_MAP __lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase = '''left''' def __init__( self :Tuple , lowercase_ :Optional[Any] , lowercase_ :Optional[int]=False , lowercase_ :int=True , lowercase_ :List[Any]=False , lowercase_ :int="<s>" , lowercase_ :Union[str, Any]="</s>" , lowercase_ :List[str]="<unk>" , lowercase_ :str="<sep>" , lowercase_ :Optional[Any]="<pad>" , lowercase_ :Tuple="<cls>" , lowercase_ :int="<mask>" , lowercase_ :Union[str, Any]=["<eop>", "<eod>"] , lowercase_ :Optional[Dict[str, Any]] = None , **lowercase_ :int , )-> Optional[int]: A__ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else mask_token A__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_snake_case , remove_space=_snake_case , keep_accents=_snake_case , bos_token=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , additional_special_tokens=_snake_case , sp_model_kwargs=self.sp_model_kwargs , **_snake_case , ) A__ = 3 A__ = do_lower_case A__ = remove_space A__ = keep_accents A__ = vocab_file A__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_snake_case ) @property def UpperCAmelCase_ ( self :Tuple )-> Any: return len(self.sp_model ) def UpperCAmelCase_ ( self :Union[str, Any] )-> List[Any]: A__ = {self.convert_ids_to_tokens(_snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self :Any )-> Union[str, Any]: A__ = self.__dict__.copy() A__ = None return state def __setstate__( self :str , lowercase_ :Union[str, Any] )-> List[str]: A__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A__ = {} A__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase_ ( self :int , lowercase_ :Any )-> Any: if self.remove_space: A__ = " ".join(inputs.strip().split() ) else: A__ = inputs A__ = outputs.replace("``" , "\"" ).replace("\'\'" , "\"" ) if not self.keep_accents: A__ = unicodedata.normalize("NFKD" , _snake_case ) A__ = "".join([c for c in outputs if not unicodedata.combining(_snake_case )] ) if self.do_lower_case: A__ = outputs.lower() return outputs def UpperCAmelCase_ ( self :Any , lowercase_ :str )-> Tuple: A__ = self.preprocess_text(_snake_case ) A__ = self.sp_model.encode(_snake_case , out_type=_snake_case ) A__ = [] for piece in pieces: if len(_snake_case ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): A__ = self.sp_model.EncodeAsPieces(piece[:-1].replace(_snake_case , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ = cur_pieces[1:] else: A__ = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_snake_case ) else: new_pieces.append(_snake_case ) return new_pieces def UpperCAmelCase_ ( self :Tuple , lowercase_ :Dict )-> Dict: return self.sp_model.PieceToId(_snake_case ) def UpperCAmelCase_ ( self :str , lowercase_ :Union[str, Any] )-> List[Any]: return self.sp_model.IdToPiece(_snake_case ) def UpperCAmelCase_ ( self :List[Any] , lowercase_ :Optional[Any] )-> Dict: A__ = "".join(_snake_case ).replace(_snake_case , " " ).strip() return out_string def UpperCAmelCase_ ( self :Any , lowercase_ :List[int] , lowercase_ :bool = False , lowercase_ :bool = None , lowercase_ :bool = True , **lowercase_ :Dict , )-> Optional[Any]: A__ = kwargs.pop("use_source_tokenizer" , _snake_case ) A__ = self.convert_ids_to_tokens(_snake_case , skip_special_tokens=_snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ = [] A__ = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_snake_case ) ) A__ = [] sub_texts.append(_snake_case ) else: current_sub_text.append(_snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ = "".join(_snake_case ) A__ = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ = self.clean_up_tokenization(_snake_case ) return clean_text else: return text def UpperCAmelCase_ ( self :Dict , lowercase_ :List[int] , lowercase_ :Optional[List[int]] = None )-> Optional[int]: A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase_ ( self :Tuple , lowercase_ :List[int] , lowercase_ :Optional[List[int]] = None , lowercase_ :bool = False )-> int: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_snake_case , token_ids_a=_snake_case , already_has_special_tokens=_snake_case ) if token_ids_a is not None: return ([0] * len(_snake_case )) + [1] + ([0] * len(_snake_case )) + [1, 1] return ([0] * len(_snake_case )) + [1, 1] def UpperCAmelCase_ ( self :Any , lowercase_ :List[int] , lowercase_ :Optional[List[int]] = None )-> int: A__ = [self.sep_token_id] A__ = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCAmelCase_ ( self :Tuple , lowercase_ :str , lowercase_ :Optional[str] = None )-> Optional[int]: if not os.path.isdir(_snake_case ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return A__ = os.path.join( _snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _snake_case ) elif not os.path.isfile(self.vocab_file ): with open(_snake_case , "wb" ) as fi: A__ = self.sp_model.serialized_model_proto() fi.write(_snake_case ) return (out_vocab_file,)
357
'''simple docstring''' import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def UpperCamelCase ( ): print("Making key files..." ) make_key_files("rsa" , 10_24 ) print("Key files generation successful." ) def UpperCamelCase ( _lowerCamelCase : int ): print("Generating prime p..." ) A__ = rabinMiller.generate_large_prime(_lowerCamelCase ) print("Generating prime q..." ) A__ = rabinMiller.generate_large_prime(_lowerCamelCase ) A__ = p * q print("Generating e that is relatively prime to (p - 1) * (q - 1)..." ) while True: A__ = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(_lowerCamelCase , (p - 1) * (q - 1) ) == 1: break print("Calculating d that is mod inverse of e..." ) A__ = cryptoMath.find_mod_inverse(_lowerCamelCase , (p - 1) * (q - 1) ) A__ = (n, e) A__ = (n, d) return (public_key, private_key) def UpperCamelCase ( _lowerCamelCase : str , _lowerCamelCase : int ): if os.path.exists(F"{name}_pubkey.txt" ) or os.path.exists(F"{name}_privkey.txt" ): print("\nWARNING:" ) print( F"\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n" "Use a different name or delete these files and re-run this program." ) sys.exit() A__, A__ = generate_key(_lowerCamelCase ) print(F"\nWriting public key to file {name}_pubkey.txt..." ) with open(F"{name}_pubkey.txt" , "w" ) as out_file: out_file.write(F"{key_size},{public_key[0]},{public_key[1]}" ) print(F"Writing private key to file {name}_privkey.txt..." ) with open(F"{name}_privkey.txt" , "w" ) as out_file: out_file.write(F"{key_size},{private_key[0]},{private_key[1]}" ) if __name__ == "__main__": main()
123
0
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def _a ( lowerCamelCase: int , lowerCamelCase: int , lowerCamelCase: int , lowerCamelCase: int , lowerCamelCase: int , lowerCamelCase: int ) -> np.ndarray: '''simple docstring''' if (ksize % 2) == 0: __A = ksize + 1 __A = np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(lowerCamelCase ): for x in range(lowerCamelCase ): # distance from center __A = x - ksize // 2 __A = y - ksize // 2 # degree to radiant __A = theta / 1_80 * np.pi __A = np.cos(_theta ) __A = np.sin(_theta ) # get kernel x __A = cos_theta * px + sin_theta * py # get kernel y __A = -sin_theta * px + cos_theta * py # fill kernel __A = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image snake_case__ : List[str] = imread('../image_data/lena.jpg') # turn image in gray scale value snake_case__ : Union[str, Any] = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges snake_case__ : str = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 120, 150]: snake_case__ : Dict = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) snake_case__ : str = out / out.max() * 255 snake_case__ : Dict = out.astype(np.uinta) imshow('Original', gray) imshow('Gabor filter with 20x20 mask and 6 directions', out) waitKey(0)
117
import argparse import copy def _a ( lowerCamelCase: List[Any] ) -> List[str]: '''simple docstring''' __A = {} with open(lowerCamelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __A = [] _list.append([line.split()[1], line.split()[2]] ) __A = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __A = [] _list.append([line.split()[0], line.split()[2]] ) __A = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _a ( lowerCamelCase: Any , lowerCamelCase: Optional[Any] ) -> Dict: '''simple docstring''' with open(lowerCamelCase ) as f: __A = f.read(1 ) __A = start_node __A = [] __A = start_node __A = 0 while visiting not in first_solution: __A = 1_00_00 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCamelCase ) and k[0] not in first_solution: __A = k[1] __A = k[0] first_solution.append(lowerCamelCase ) __A = distance_of_first_solution + int(lowerCamelCase ) __A = best_node first_solution.append(lowerCamelCase ) __A = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __A = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_00_00 ) return first_solution, distance_of_first_solution def _a ( lowerCamelCase: List[str] , lowerCamelCase: Any ) -> Any: '''simple docstring''' __A = [] for n in solution[1:-1]: __A = solution.index(lowerCamelCase ) for kn in solution[1:-1]: __A = solution.index(lowerCamelCase ) if n == kn: continue __A = copy.deepcopy(lowerCamelCase ) __A = kn __A = n __A = 0 for k in _tmp[:-1]: __A = _tmp[_tmp.index(lowerCamelCase ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __A = distance + int(i[1] ) _tmp.append(lowerCamelCase ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __A = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowerCamelCase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _a ( lowerCamelCase: Optional[int] , lowerCamelCase: Dict , lowerCamelCase: Any , lowerCamelCase: Optional[int] , lowerCamelCase: Union[str, Any] ) -> Any: '''simple docstring''' __A = 1 __A = first_solution __A = [] __A = distance_of_first_solution __A = solution while count <= iters: __A = find_neighborhood(lowerCamelCase , lowerCamelCase ) __A = 0 __A = neighborhood[index_of_best_solution] __A = len(lowerCamelCase ) - 1 __A = False while not found: __A = 0 while i < len(lowerCamelCase ): if best_solution[i] != solution[i]: __A = best_solution[i] __A = solution[i] break __A = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __A = True __A = best_solution[:-1] __A = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __A = cost __A = solution else: __A = index_of_best_solution + 1 __A = neighborhood[index_of_best_solution] if len(lowerCamelCase ) >= size: tabu_list.pop(0 ) __A = count + 1 return best_solution_ever, best_cost def _a ( lowerCamelCase: List[str]=None ) -> str: '''simple docstring''' __A = generate_neighbours(args.File ) __A , __A = generate_first_solution( args.File , lowerCamelCase ) __A , __A = tabu_search( lowerCamelCase , lowerCamelCase , lowerCamelCase , args.Iterations , args.Size , ) print(F"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": snake_case__ : Tuple = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
117
1
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument( """--original_config_file""", type=str, required=True, help="""The YAML config file corresponding to the original architecture.""", ) parser.add_argument( """--num_in_channels""", default=None, type=int, help="""The number of input channels. If `None` number of input channels will be automatically inferred.""", ) parser.add_argument( """--image_size""", default=512, type=int, help=( """The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2""" """ Base. Use 768 for Stable Diffusion v2.""" ), ) parser.add_argument( """--extract_ema""", action="""store_true""", help=( """Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights""" """ or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield""" """ higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.""" ), ) parser.add_argument( """--upcast_attention""", action="""store_true""", help=( """Whether the attention computation should always be upcasted. This is necessary when running stable""" """ diffusion 2.1.""" ), ) parser.add_argument( """--from_safetensors""", action="""store_true""", help="""If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.""", ) parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") def SCREAMING_SNAKE_CASE__ ( snake_case : int )-> Tuple: '''simple docstring''' if string == "True": return True elif string == "False": return False else: raise ValueError(f'could not parse string as bool {string}' ) parser.add_argument( """--use_linear_projection""", help="""Override for use linear projection""", required=False, type=parse_bool ) parser.add_argument("""--cross_attention_dim""", help="""Override for cross attention_dim""", required=False, type=int) _lowerCAmelCase : Optional[Any] = parser.parse_args() _lowerCAmelCase : Optional[Any] = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
369
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class lowerCAmelCase__ : def __init__( self : Optional[int] , snake_case__ : List[Any] , snake_case__ : str=sys.maxsize ): '''simple docstring''' UpperCAmelCase__ : Any = "bilinear" UpperCAmelCase__ : Any = max_size UpperCAmelCase__ : Any = short_edge_length def __call__( self : Dict , snake_case__ : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = [] for img in imgs: UpperCAmelCase__ , UpperCAmelCase__ : int = img.shape[:2] # later: provide list and randomly choose index for resize UpperCAmelCase__ : Dict = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img UpperCAmelCase__ : Dict = size * 1.0 / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = size, scale * w else: UpperCAmelCase__ , UpperCAmelCase__ : int = scale * h, size if max(snake_case__ , snake_case__ ) > self.max_size: UpperCAmelCase__ : Union[str, Any] = self.max_size * 1.0 / max(snake_case__ , snake_case__ ) UpperCAmelCase__ : List[str] = newh * scale UpperCAmelCase__ : int = neww * scale UpperCAmelCase__ : List[Any] = int(neww + 0.5 ) UpperCAmelCase__ : Optional[Any] = int(newh + 0.5 ) if img.dtype == np.uinta: UpperCAmelCase__ : Any = Image.fromarray(snake_case__ ) UpperCAmelCase__ : Union[str, Any] = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) UpperCAmelCase__ : Optional[int] = np.asarray(snake_case__ ) else: UpperCAmelCase__ : Any = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw UpperCAmelCase__ : Tuple = nn.functional.interpolate( snake_case__ , (newh, neww) , mode=self.interp_method , align_corners=snake_case__ ).squeeze(0 ) img_augs.append(snake_case__ ) return img_augs class lowerCAmelCase__ : def __init__( self : Optional[int] , snake_case__ : Dict ): '''simple docstring''' UpperCAmelCase__ : Dict = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) UpperCAmelCase__ : Any = cfg.INPUT.FORMAT UpperCAmelCase__ : Optional[Any] = cfg.SIZE_DIVISIBILITY UpperCAmelCase__ : str = cfg.PAD_VALUE UpperCAmelCase__ : List[Any] = cfg.INPUT.MAX_SIZE_TEST UpperCAmelCase__ : Dict = cfg.MODEL.DEVICE UpperCAmelCase__ : Optional[int] = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase__ : str = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase__ : List[str] = lambda snake_case__ : (x - self.pixel_mean) / self.pixel_std def __a ( self : Optional[int] , snake_case__ : Dict ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = tuple(max(snake_case__ ) for s in zip(*[img.shape for img in images] ) ) UpperCAmelCase__ : Tuple = [im.shape[-2:] for im in images] UpperCAmelCase__ : int = [ nn.functional.pad( snake_case__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(snake_case__ , snake_case__ ) ] return torch.stack(snake_case__ ), torch.tensor(snake_case__ ) def __call__( self : str , snake_case__ : int , snake_case__ : int=False ): '''simple docstring''' with torch.no_grad(): if not isinstance(snake_case__ , snake_case__ ): UpperCAmelCase__ : Dict = [images] if single_image: assert len(snake_case__ ) == 1 for i in range(len(snake_case__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(snake_case__ , images.pop(snake_case__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( snake_case__ , torch.as_tensor(img_tensorize(images.pop(snake_case__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge UpperCAmelCase__ : Optional[Any] = torch.tensor([im.shape[:2] for im in images] ) UpperCAmelCase__ : Tuple = self.aug(snake_case__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic UpperCAmelCase__ : Optional[int] = [self.normalizer(snake_case__ ) for x in images] # now pad them to do the following operations UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.pad(snake_case__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad UpperCAmelCase__ : Tuple = torch.true_divide(snake_case__ , snake_case__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def SCREAMING_SNAKE_CASE__ ( snake_case : Optional[Any] , snake_case : str )-> List[Any]: '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def SCREAMING_SNAKE_CASE__ ( snake_case : List[Any] , snake_case : Tuple[int, int] )-> int: '''simple docstring''' assert torch.isfinite(snake_case ).all(), "Box tensor contains infinite or NaN!" UpperCAmelCase__ , UpperCAmelCase__ : Dict = box_size tensor[:, 0].clamp_(min=0 , max=snake_case ) tensor[:, 1].clamp_(min=0 , max=snake_case ) tensor[:, 2].clamp_(min=0 , max=snake_case ) tensor[:, 3].clamp_(min=0 , max=snake_case )
298
0
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __a: str = 16 __a: List[Any] = 32 def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase = 16 ): lowercase__ : Union[str, Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowercase__ : List[str] = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) lowercase__ : List[Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__UpperCamelCase , max_length=__UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase__ : str = datasets.map( __UpperCamelCase , batched=__UpperCamelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__ : Dict = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase__ : List[str] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase__ : Optional[Any] = 16 elif accelerator.mixed_precision != "no": lowercase__ : Optional[Any] = 8 else: lowercase__ : Tuple = None return tokenizer.pad( __UpperCamelCase , padding='''longest''' , max_length=__UpperCamelCase , pad_to_multiple_of=__UpperCamelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. lowercase__ : Tuple = DataLoader( tokenized_datasets['''train'''] , shuffle=__UpperCamelCase , collate_fn=__UpperCamelCase , batch_size=__UpperCamelCase ) lowercase__ : Optional[Any] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__UpperCamelCase , collate_fn=__UpperCamelCase , batch_size=__UpperCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders __a: Tuple = mocked_dataloaders # noqa: F811 def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __UpperCamelCase ) == "1": lowercase__ : List[Any] = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: lowercase__ : Optional[int] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with='''all''' , project_dir=args.project_dir ) else: lowercase__ : List[str] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__ : int = config['''lr'''] lowercase__ : str = int(config['''num_epochs'''] ) lowercase__ : Any = int(config['''seed'''] ) lowercase__ : Union[str, Any] = int(config['''batch_size'''] ) set_seed(__UpperCamelCase ) lowercase__ , lowercase__ : Tuple = get_dataloaders(__UpperCamelCase , __UpperCamelCase ) lowercase__ : Any = evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation lowercase__ : Tuple = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowercase__ : Tuple = batch_size // MAX_GPU_BATCH_SIZE lowercase__ : Optional[int] = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__ : Optional[Any] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__UpperCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase__ : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer lowercase__ : Union[str, Any] = AdamW(params=model.parameters() , lr=__UpperCamelCase ) # Instantiate scheduler lowercase__ : List[str] = get_linear_schedule_with_warmup( optimizer=__UpperCamelCase , num_warmup_steps=100 , num_training_steps=(len(__UpperCamelCase ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Optional[int] = accelerator.prepare( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: lowercase__ : Tuple = os.path.split(__UpperCamelCase )[-1].split('''.''' )[0] accelerator.init_trackers(__UpperCamelCase , __UpperCamelCase ) # Now we train the model for epoch in range(__UpperCamelCase ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: lowercase__ : Any = 0 for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowercase__ : Any = model(**__UpperCamelCase ) lowercase__ : int = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() lowercase__ : Tuple = loss / gradient_accumulation_steps accelerator.backward(__UpperCamelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): lowercase__ : List[Any] = model(**__UpperCamelCase ) lowercase__ : str = outputs.logits.argmax(dim=-1 ) lowercase__ , lowercase__ : List[Any] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__UpperCamelCase , references=__UpperCamelCase , ) lowercase__ : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , __UpperCamelCase ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { '''accuracy''': eval_metric['''accuracy'''], '''f1''': eval_metric['''f1'''], '''train_loss''': total_loss.item() / len(__UpperCamelCase ), '''epoch''': epoch, } , step=__UpperCamelCase , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def __UpperCamelCase ( ): lowercase__ : Tuple = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__UpperCamelCase , default=__UpperCamelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) parser.add_argument( '''--with_tracking''' , action='''store_true''' , help='''Whether to load in all available experiment trackers from the environment and use them for logging.''' , ) parser.add_argument( '''--project_dir''' , type=__UpperCamelCase , default='''logs''' , help='''Location on where to store experiment tracking logs` and relevent project information''' , ) lowercase__ : Tuple = parser.parse_args() lowercase__ : Optional[int] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__UpperCamelCase , __UpperCamelCase ) if __name__ == "__main__": main()
198
import math from datetime import datetime, timedelta def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = year % 1_9 SCREAMING_SNAKE_CASE_ = year % 4 SCREAMING_SNAKE_CASE_ = year % 7 SCREAMING_SNAKE_CASE_ = math.floor(year / 1_0_0 ) SCREAMING_SNAKE_CASE_ = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) SCREAMING_SNAKE_CASE_ = leap_day_inhibits / 4 SCREAMING_SNAKE_CASE_ = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 SCREAMING_SNAKE_CASE_ = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 SCREAMING_SNAKE_CASE_ = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon SCREAMING_SNAKE_CASE_ = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(__UpperCamelCase , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(__UpperCamelCase , 4 , 1_8 ) else: return datetime(__UpperCamelCase , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): A : Dict = "will be" if year > datetime.now().year else "was" print(f"Easter in {year} {tense} {gauss_easter(year)}")
118
0
import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class __lowercase (unittest.TestCase ): """simple docstring""" _UpperCAmelCase = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _UpperCAmelCase = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = TextaTextGenerationPipeline(model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) return generator, ["Something to write", "Something else"] def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = generator('Something there' ) self.assertEqual(lowerCAmelCase__ , [{'generated_text': ANY(lowerCAmelCase__ )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['generated_text'].startswith('Something there' ) ) SCREAMING_SNAKE_CASE_ : List[Any] = generator(['This is great !', 'Something else'] , num_return_sequences=2 , do_sample=lowerCAmelCase__ ) self.assertEqual( lowerCAmelCase__ , [ [{'generated_text': ANY(lowerCAmelCase__ )}, {'generated_text': ANY(lowerCAmelCase__ )}], [{'generated_text': ANY(lowerCAmelCase__ )}, {'generated_text': ANY(lowerCAmelCase__ )}], ] , ) SCREAMING_SNAKE_CASE_ : Dict = generator( ['This is great !', 'Something else'] , num_return_sequences=2 , batch_size=2 , do_sample=lowerCAmelCase__ ) self.assertEqual( lowerCAmelCase__ , [ [{'generated_text': ANY(lowerCAmelCase__ )}, {'generated_text': ANY(lowerCAmelCase__ )}], [{'generated_text': ANY(lowerCAmelCase__ )}, {'generated_text': ANY(lowerCAmelCase__ )}], ] , ) with self.assertRaises(lowerCAmelCase__ ): generator(4 ) @require_torch def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='pt' ) # do_sample=False necessary for reproducibility SCREAMING_SNAKE_CASE_ : Optional[int] = generator('Something there' , do_sample=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , [{'generated_text': ''}] ) SCREAMING_SNAKE_CASE_ : Any = 3 SCREAMING_SNAKE_CASE_ : Dict = generator( 'Something there' , num_return_sequences=lowerCAmelCase__ , num_beams=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_ : Optional[int] = [ {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': ''}, ] self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = generator('This is a test' , do_sample=lowerCAmelCase__ , num_return_sequences=2 , return_tensors=lowerCAmelCase__ ) self.assertEqual( lowerCAmelCase__ , [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ] , ) SCREAMING_SNAKE_CASE_ : List[str] = generator.model.config.eos_token_id SCREAMING_SNAKE_CASE_ : Tuple = '<pad>' SCREAMING_SNAKE_CASE_ : Optional[int] = generator( ['This is a test', 'This is a second test'] , do_sample=lowerCAmelCase__ , num_return_sequences=2 , batch_size=2 , return_tensors=lowerCAmelCase__ , ) self.assertEqual( lowerCAmelCase__ , [ [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], ] , ) @require_tf def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='tf' ) # do_sample=False necessary for reproducibility SCREAMING_SNAKE_CASE_ : int = generator('Something there' , do_sample=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , [{'generated_text': ''}] )
162
import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def a__ ( A__, A__, A__, A__, A__ ): # Load configuration defined in the metadata file with open(A__ ) as metadata_file: SCREAMING_SNAKE_CASE_ : List[str] = json.load(A__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = LukeConfig(use_entity_aware_attention=A__, **metadata['model_config'] ) # Load in the weights from the checkpoint_path SCREAMING_SNAKE_CASE_ : Optional[int] = torch.load(A__, map_location='cpu' )['module'] # Load the entity vocab file SCREAMING_SNAKE_CASE_ : Union[str, Any] = load_original_entity_vocab(A__ ) # add an entry for [MASK2] SCREAMING_SNAKE_CASE_ : str = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 SCREAMING_SNAKE_CASE_ : Union[str, Any] = XLMRobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks SCREAMING_SNAKE_CASE_ : Any = AddedToken('<ent>', lstrip=A__, rstrip=A__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = AddedToken('<ent2>', lstrip=A__, rstrip=A__ ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(A__ ) with open(os.path.join(A__, 'tokenizer_config.json' ), 'r' ) as f: SCREAMING_SNAKE_CASE_ : str = json.load(A__ ) SCREAMING_SNAKE_CASE_ : List[Any] = 'MLukeTokenizer' with open(os.path.join(A__, 'tokenizer_config.json' ), 'w' ) as f: json.dump(A__, A__ ) with open(os.path.join(A__, MLukeTokenizer.vocab_files_names['entity_vocab_file'] ), 'w' ) as f: json.dump(A__, A__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = MLukeTokenizer.from_pretrained(A__ ) # Initialize the embeddings of the special tokens SCREAMING_SNAKE_CASE_ : Tuple = tokenizer.convert_tokens_to_ids(['@'] )[0] SCREAMING_SNAKE_CASE_ : str = tokenizer.convert_tokens_to_ids(['#'] )[0] SCREAMING_SNAKE_CASE_ : str = state_dict['embeddings.word_embeddings.weight'] SCREAMING_SNAKE_CASE_ : int = word_emb[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : List[str] = word_emb[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : List[str] = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: SCREAMING_SNAKE_CASE_ : Tuple = state_dict[bias_name] SCREAMING_SNAKE_CASE_ : Any = decoder_bias[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = decoder_bias[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : List[Any] = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: SCREAMING_SNAKE_CASE_ : Tuple = F'''encoder.layer.{layer_index}.attention.self.''' SCREAMING_SNAKE_CASE_ : Tuple = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE_ : Dict = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE_ : Optional[Any] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks SCREAMING_SNAKE_CASE_ : Union[str, Any] = state_dict['entity_embeddings.entity_embeddings.weight'] SCREAMING_SNAKE_CASE_ : Union[str, Any] = entity_emb[entity_vocab['[MASK]']].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : List[str] = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' SCREAMING_SNAKE_CASE_ : List[str] = state_dict['entity_predictions.bias'] SCREAMING_SNAKE_CASE_ : str = entity_prediction_bias[entity_vocab['[MASK]']].unsqueeze(0 ) SCREAMING_SNAKE_CASE_ : Optional[Any] = torch.cat([entity_prediction_bias, entity_mask_bias] ) SCREAMING_SNAKE_CASE_ : Tuple = LukeForMaskedLM(config=A__ ).eval() state_dict.pop('entity_predictions.decoder.weight' ) state_dict.pop('lm_head.decoder.weight' ) state_dict.pop('lm_head.decoder.bias' ) SCREAMING_SNAKE_CASE_ : str = OrderedDict() for key, value in state_dict.items(): if not (key.startswith('lm_head' ) or key.startswith('entity_predictions' )): SCREAMING_SNAKE_CASE_ : str = state_dict[key] else: SCREAMING_SNAKE_CASE_ : Dict = state_dict[key] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = model.load_state_dict(A__, strict=A__ ) if set(A__ ) != {"luke.embeddings.position_ids"}: raise ValueError(F'''Unexpected unexpected_keys: {unexpected_keys}''' ) if set(A__ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F'''Unexpected missing_keys: {missing_keys}''' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs SCREAMING_SNAKE_CASE_ : List[str] = MLukeTokenizer.from_pretrained(A__, task='entity_classification' ) SCREAMING_SNAKE_CASE_ : Any = 'ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).' SCREAMING_SNAKE_CASE_ : Dict = (0, 9) SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer(A__, entity_spans=[span], return_tensors='pt' ) SCREAMING_SNAKE_CASE_ : List[Any] = model(**A__ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE_ : List[str] = torch.Size((1, 3_3, 7_6_8) ) SCREAMING_SNAKE_CASE_ : Tuple = torch.tensor([[0.08_92, 0.05_96, -0.28_19], [0.01_34, 0.11_99, 0.05_73], [-0.01_69, 0.09_27, 0.06_44]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3], A__, atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE_ : Dict = torch.Size((1, 1, 7_6_8) ) SCREAMING_SNAKE_CASE_ : int = torch.tensor([[-0.14_82, 0.06_09, 0.03_22]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3], A__, atol=1E-4 ): raise ValueError # Verify masked word/entity prediction SCREAMING_SNAKE_CASE_ : Optional[int] = MLukeTokenizer.from_pretrained(A__ ) SCREAMING_SNAKE_CASE_ : Tuple = 'Tokyo is the capital of <mask>.' SCREAMING_SNAKE_CASE_ : Tuple = (2_4, 3_0) SCREAMING_SNAKE_CASE_ : List[str] = tokenizer(A__, entity_spans=[span], return_tensors='pt' ) SCREAMING_SNAKE_CASE_ : Tuple = model(**A__ ) SCREAMING_SNAKE_CASE_ : List[Any] = encoding['input_ids'][0].tolist() SCREAMING_SNAKE_CASE_ : Union[str, Any] = input_ids.index(tokenizer.convert_tokens_to_ids('<mask>' ) ) SCREAMING_SNAKE_CASE_ : Any = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(A__ ) SCREAMING_SNAKE_CASE_ : int = outputs.entity_logits[0][0].argmax().item() SCREAMING_SNAKE_CASE_ : List[str] = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith('en:' )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(A__ ) ) model.save_pretrained(A__ ) def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Tuple = ['[MASK]', '[PAD]', '[UNK]'] SCREAMING_SNAKE_CASE_ : int = [json.loads(A__ ) for line in open(A__ )] SCREAMING_SNAKE_CASE_ : Union[str, Any] = {} for entry in data: SCREAMING_SNAKE_CASE_ : List[Any] = entry['id'] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: SCREAMING_SNAKE_CASE_ : List[Any] = entity_id break SCREAMING_SNAKE_CASE_ : int = F'''{language}:{entity_name}''' SCREAMING_SNAKE_CASE_ : Optional[int] = entity_id return new_mapping if __name__ == "__main__": lowerCAmelCase__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Path to a pytorch_model.bin file.') parser.add_argument( '--metadata_path', default=None, type=str, help='Path to a metadata.json file, defining the configuration.' ) parser.add_argument( '--entity_vocab_path', default=None, type=str, help='Path to an entity_vocab.tsv file, containing the entity vocabulary.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to where to dump the output PyTorch model.' ) parser.add_argument( '--model_size', default='base', type=str, choices=['base', 'large'], help='Size of the model to be converted.' ) lowerCAmelCase__ : List[Any] =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
162
1
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient A =WebClient(token=os.environ['CI_SLACK_BOT_TOKEN']) def snake_case_ (_a : Tuple ): UpperCAmelCase = test_results.split(''' ''' ) UpperCAmelCase = 0 UpperCAmelCase = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. UpperCAmelCase = expressions[-2] if '''=''' in expressions[-1] else expressions[-1] for i, expression in enumerate(_a ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def snake_case_ (_a : Optional[int] ): UpperCAmelCase = {} UpperCAmelCase = None UpperCAmelCase = False for line in failures_short_lines.split('''\n''' ): if re.search(R'''_ \[doctest\]''' , _a ): UpperCAmelCase = True UpperCAmelCase = line.split(''' ''' )[2] elif in_error and not line.split(''' ''' )[0].isdigit(): UpperCAmelCase = line UpperCAmelCase = False return failures class _a : def __init__( self : Dict , lowercase : str , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = title UpperCAmelCase = doc_test_results['''time_spent'''].split(''',''' )[0] UpperCAmelCase = doc_test_results['''success'''] UpperCAmelCase = doc_test_results['''failures'''] UpperCAmelCase = self.n_success + self.n_failures # Failures and success of the modeling tests UpperCAmelCase = doc_test_results @property def A ( self : Any ): '''simple docstring''' UpperCAmelCase = [self._time_spent] UpperCAmelCase = 0 for time in time_spent: UpperCAmelCase = time.split(''':''' ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(lowercase ) == 1: UpperCAmelCase = [0, 0, time_parts[0]] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3_600 + minutes * 60 + seconds UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = total_secs // 3_600, (total_secs % 3_600) // 60, total_secs % 60 return f"{int(lowercase )}h{int(lowercase )}m{int(lowercase )}s" @property def A ( self : int ): '''simple docstring''' return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def A ( self : Union[str, Any] ): '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": f"🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.", "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def A ( self : int ): '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": ( f"There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in" f" {self.time}." ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = 40 UpperCAmelCase = {k: v['''failed'''] for k, v in doc_test_results.items() if isinstance(lowercase , lowercase )} UpperCAmelCase = '''''' for category, failures in category_failures.items(): if len(lowercase ) == 0: continue if report != "": report += "\n\n" report += f"*{category} failures*:".ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(lowercase ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f"The following examples had failures:\n\n\n{report}\n", }, } @property def A ( self : Any ): '''simple docstring''' UpperCAmelCase = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(lowercase ) @staticmethod def A ( ): '''simple docstring''' UpperCAmelCase = [ { '''type''': '''section''', '''text''': { '''type''': '''plain_text''', '''text''': '''There was an issue running the tests.''', }, '''accessory''': { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''Check Action results''', '''emoji''': True}, '''url''': f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } ] print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(lowercase )} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , text='''There was an issue running the tests.''' , blocks=lowercase , ) def A ( self : Optional[Any] ): '''simple docstring''' print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(self.payload )} ) ) UpperCAmelCase = f"{self.n_failures} failures out of {self.n_tests} tests," if self.n_failures else '''All tests passed.''' UpperCAmelCase = client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , blocks=self.payload , text=lowercase , ) def A ( self : Optional[int] , lowercase : Any , lowercase : Any , lowercase : Union[str, Any] , lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = '''''' for key, value in failures.items(): UpperCAmelCase = value[:200] + ''' [Truncated]''' if len(lowercase ) > 250 else value failures_text += f"*{key}*\n_{value}_\n\n" UpperCAmelCase = job_name UpperCAmelCase = {'''type''': '''section''', '''text''': {'''type''': '''mrkdwn''', '''text''': text}} if job_link is not None: UpperCAmelCase = { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''GitHub Action job''', '''emoji''': True}, '''url''': job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def A ( self : Optional[int] ): '''simple docstring''' if self.thread_ts is None: raise ValueError('''Can only post reply if a post has been made.''' ) UpperCAmelCase = self.doc_test_results.pop('''job_link''' ) self.doc_test_results.pop('''failures''' ) self.doc_test_results.pop('''success''' ) self.doc_test_results.pop('''time_spent''' ) UpperCAmelCase = sorted(self.doc_test_results.items() , key=lambda lowercase : t[0] ) for job, job_result in sorted_dict: if len(job_result['''failures'''] ): UpperCAmelCase = f"*Num failures* :{len(job_result['failed'] )} \n" UpperCAmelCase = job_result['''failures'''] UpperCAmelCase = self.get_reply_blocks(lowercase , lowercase , lowercase , text=lowercase ) print('''Sending the following reply''' ) print(json.dumps({'''blocks''': blocks} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] , text=f"Results for {job}" , blocks=lowercase , thread_ts=self.thread_ts['''ts'''] , ) time.sleep(1 ) def snake_case_ (): UpperCAmelCase = os.environ['''GITHUB_RUN_ID'''] UpperCAmelCase = F"https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100" UpperCAmelCase = requests.get(_a ).json() UpperCAmelCase = {} try: jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) UpperCAmelCase = math.ceil((result['''total_count'''] - 1_0_0) / 1_0_0 ) for i in range(_a ): UpperCAmelCase = requests.get(url + F"&page={i + 2}" ).json() jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) return jobs except Exception as e: print('''Unknown error, could not fetch links.''' , _a ) return {} def snake_case_ (_a : str ): UpperCAmelCase = {} if os.path.exists(_a ): UpperCAmelCase = os.listdir(_a ) for file in files: try: with open(os.path.join(_a , _a ) , encoding='''utf-8''' ) as f: UpperCAmelCase = f.read() except UnicodeDecodeError as e: raise ValueError(F"Could not open {os.path.join(_a , _a )}." ) from e return _artifact def snake_case_ (): class _a : def __init__( self : Any , lowercase : str ): '''simple docstring''' UpperCAmelCase = name UpperCAmelCase = [] def __str__( self : Tuple ): '''simple docstring''' return self.name def A ( self : List[Any] , lowercase : str ): '''simple docstring''' self.paths.append({'''name''': self.name, '''path''': path} ) UpperCAmelCase = {} UpperCAmelCase = filter(os.path.isdir , os.listdir() ) for directory in directories: UpperCAmelCase = directory if artifact_name not in _available_artifacts: UpperCAmelCase = Artifact(_a ) _available_artifacts[artifact_name].add_path(_a ) return _available_artifacts if __name__ == "__main__": A =get_job_links() A =retrieve_available_artifacts() A =collections.OrderedDict( [ ('*.py', 'API Examples'), ('*.md', 'MD Examples'), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' A ={ v: { 'failed': [], 'failures': {}, } for v in docs.values() } # Link to the GitHub Action job A =github_actions_job_links.get('run_doctests') A =available_artifacts['doc_tests_gpu_test_reports'].paths[0] A =retrieve_artifact(artifact_path['name']) if "stats" in artifact: A , A , A =handle_test_results(artifact['stats']) A =failed A =success A =time_spent[1:-1] + ', ' A =extract_first_line_failure(artifact['failures_short']) for line in artifact["summary_short"].split('\n'): if re.search('FAILED', line): A =line.replace('FAILED ', '') A =line.split()[0].replace('\n', '') if "::" in line: A , A =line.split('::') else: A , A =line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): A =docs[file_regex] doc_test_results[category]["failed"].append(test) A =all_failures[test] if test in all_failures else 'N/A' A =failure break A =Message('🤗 Results of the doc tests.', doc_test_results) message.post() message.post_reply()
34
"""simple docstring""" from scipy.stats import spearmanr import datasets _a = '\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n' _a = '\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {\'spearmanr\': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results[\'spearmanr\'])\n -0.7\n >>> print(round(results[\'spearmanr_pvalue\'], 2))\n 0.19\n' _a = r'\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def _lowercase ( self : Dict ): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ), reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"], ) def _lowercase ( self : Optional[Any], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[int]=False ): __lowercase = spearmanr(UpperCAmelCase__, UpperCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
17
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __lowerCamelCase ( lowerCAmelCase_ ) -> Tuple: _a : List[Any] = filter(lambda lowerCAmelCase_ : p.requires_grad , model.parameters() ) _a : Any = sum([np.prod(p.size() ) for p in model_parameters] ) return params __lowerCAmelCase = logging.getLogger(__name__) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> int: if metric == "rouge2": _a : Union[str, Any] = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": _a : Tuple = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": _a : Any = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) _a : Tuple = ModelCheckpoint( dirpath=lowerCAmelCase_ , filename=lowerCAmelCase_ , monitor=f"""val_{metric}""" , mode='max' , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Tuple: return EarlyStopping( monitor=f"""val_{metric}""" , mode='min' if 'loss' in metric else 'max' , patience=lowerCAmelCase_ , verbose=lowerCAmelCase_ , ) class __magic_name__ ( pl.Callback ): def __lowercase ( self : Optional[Any] ,_UpperCAmelCase : Tuple ,_UpperCAmelCase : Union[str, Any] ): _a : Optional[Any] = {F"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_UpperCAmelCase ) @rank_zero_only def __lowercase ( self : Union[str, Any] ,_UpperCAmelCase : pl.Trainer ,_UpperCAmelCase : pl.LightningModule ,_UpperCAmelCase : str ,_UpperCAmelCase : Optional[Any]=True ): logger.info(F"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) _a : Dict = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results _a : Tuple = Path(pl_module.hparams.output_dir ) if type_path == "test": _a : Dict = od / 'test_results.txt' _a : Dict = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. _a : Any = od / F"""{type_path}_results/{trainer.global_step:05d}.txt""" _a : Tuple = od / F"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=_UpperCAmelCase ) generations_file.parent.mkdir(exist_ok=_UpperCAmelCase ) with open(_UpperCAmelCase ,'a+' ) as writer: for key in sorted(_UpperCAmelCase ): if key in ["log", "progress_bar", "preds"]: continue _a : Tuple = metrics[key] if isinstance(_UpperCAmelCase ,torch.Tensor ): _a : Union[str, Any] = val.item() _a : Optional[Any] = F"""{key}: {val:.6f}\n""" writer.write(_UpperCAmelCase ) if not save_generations: return if "preds" in metrics: _a : Optional[int] = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(_UpperCAmelCase ) @rank_zero_only def __lowercase ( self : Any ,_UpperCAmelCase : List[str] ,_UpperCAmelCase : Optional[int] ): try: _a : Optional[Any] = pl_module.model.model.num_parameters() except AttributeError: _a : Any = pl_module.model.num_parameters() _a : List[Any] = count_trainable_parameters(_UpperCAmelCase ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1E6, 'grad_mp': n_trainable_pars / 1E6} ) @rank_zero_only def __lowercase ( self : Union[str, Any] ,_UpperCAmelCase : pl.Trainer ,_UpperCAmelCase : pl.LightningModule ): save_json(pl_module.metrics ,pl_module.metrics_save_path ) return self._write_logs(_UpperCAmelCase ,_UpperCAmelCase ,'test' ) @rank_zero_only def __lowercase ( self : Tuple ,_UpperCAmelCase : pl.Trainer ,_UpperCAmelCase : Any ): save_json(pl_module.metrics ,pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
367
'''simple docstring''' # Function to print upper half of diamond (pyramid) def __lowerCamelCase ( lowerCAmelCase_ ) -> int: for i in range(0 , lowerCAmelCase_ ): for _ in range(0 , n - i - 1 ): # printing spaces print(' ' , end='' ) for _ in range(0 , i + 1 ): # printing stars print('* ' , end='' ) print() def __lowerCamelCase ( lowerCAmelCase_ ) -> Union[str, Any]: for i in range(lowerCAmelCase_ , 0 , -1 ): for _ in range(lowerCAmelCase_ , 0 , -1 ): # printing stars print('* ' , end='' ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(' ' , end='' ) def __lowerCamelCase ( lowerCAmelCase_ ) -> Tuple: if n <= 0: print(' ... .... nothing printing :(' ) return floyd(lowerCAmelCase_ ) # upper half reverse_floyd(lowerCAmelCase_ ) # lower half if __name__ == "__main__": print(r'''| /\ | |- | |- |--| |\ /| |-''') print(r'''|/ \| |- |_ |_ |__| | \/ | |_''') __lowerCAmelCase = 1 while K: __lowerCAmelCase = int(input('''enter the number and , and see the magic : ''')) print() pretty_print(user_number) __lowerCAmelCase = int(input('''press 0 to exit... and 1 to continue...''')) print('''Good Bye...''')
107
0
def snake_case ( snake_case__ :Optional[Any]) -> Any: _A = 0 # if input_string is "aba" than new_input_string become "a|b|a" _A = '' _A = '' # append each character + "|" in new_string for range(0, length-1) for i in input_string[: len(__a) - 1]: new_input_string += i + "|" # append last character new_input_string += input_string[-1] # we will store the starting and ending of previous furthest ending palindromic # substring _A = 0, 0 # length[i] shows the length of palindromic substring with center i _A = [1 for i in range(len(__a))] # for each character in new_string find corresponding palindromic string _A = 0 for j in range(len(__a)): _A = 1 if j > r else min(length[l + r - j] // 2 , r - j + 1) while ( j - k >= 0 and j + k < len(__a) and new_input_string[k + j] == new_input_string[j - k] ): k += 1 _A = 2 * k - 1 # does this string is ending after the previously explored end (that is r) ? # if yes the update the new r to the last index of this if j + k - 1 > r: _A = j - k + 1 # noqa: E741 _A = j + k - 1 # update max_length and start position if max_length < length[j]: _A = length[j] _A = j # create that string _A = new_input_string[start - max_length // 2 : start + max_length // 2 + 1] for i in s: if i != "|": output_string += i return output_string if __name__ == "__main__": import doctest doctest.testmod()
180
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _SCREAMING_SNAKE_CASE = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """PoolFormerOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["""PoolFormerFeatureExtractor"""] _SCREAMING_SNAKE_CASE = ["""PoolFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ """POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PoolFormerForImageClassification""", """PoolFormerModel""", """PoolFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
327
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) class __lowerCAmelCase ( __a ): def __init__(self , *lowerCAmelCase__ , **lowerCAmelCase__ ): warnings.warn( """The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use SegformerImageProcessor instead.""" , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
170
'''simple docstring''' def __A ( lowerCAmelCase_ ): _UpperCAmelCase : Optional[Any] = 0 while len(lowerCAmelCase_ ) > 1: _UpperCAmelCase : List[Any] = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): _UpperCAmelCase : Optional[Any] = files.index(min(lowerCAmelCase_ ) ) temp += files[min_index] files.pop(lowerCAmelCase_ ) files.append(lowerCAmelCase_ ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
170
1
import socket def __SCREAMING_SNAKE_CASE ( ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) SCREAMING_SNAKE_CASE__ = socket.gethostname() SCREAMING_SNAKE_CASE__ = 1_23_12 sock.connect((host, port) ) sock.send(B"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: SCREAMING_SNAKE_CASE__ = sock.recv(10_24 ) if not data: break out_file.write(__UpperCamelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
219
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def __SCREAMING_SNAKE_CASE ( ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = ArgumentParser( description=( """PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes""" ) ) # Optional arguments for the launch helper parser.add_argument("""--num_cores""" , type=__UpperCamelCase , default=1 , help="""Number of TPU cores to use (1 or 8).""" ) # positional parser.add_argument( """training_script""" , type=__UpperCamelCase , help=( """The full path to the single TPU training """ """program/script to be launched in parallel, """ """followed by all the arguments for the """ """training script""" ) , ) # rest from the training program parser.add_argument("""training_script_args""" , nargs=__UpperCamelCase ) return parser.parse_args() def __SCREAMING_SNAKE_CASE ( ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = parse_args() # Import training_script as a module. SCREAMING_SNAKE_CASE__ = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) SCREAMING_SNAKE_CASE__ = script_fpath.stem SCREAMING_SNAKE_CASE__ = importlib.import_module(__UpperCamelCase ) # Patch sys.argv SCREAMING_SNAKE_CASE__ = [args.training_script] + args.training_script_args + ["""--tpu_num_cores""", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
219
1
"""simple docstring""" lowercase__ = 8.314462 # Unit - J mol-1 K-1 def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if moles < 0 or kelvin < 0 or volume < 0: raise ValueError('Invalid inputs. Enter positive value.' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError('Invalid inputs. Enter positive value.' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
365
"""simple docstring""" import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """AutoTokenizer""" lowerCamelCase__ = ["""tokenizer"""] lowerCamelCase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self , lowercase , lowercase=None ): super().__init__(lowercase ) _lowerCamelCase : Optional[int] = speaker_embeddings @classmethod def A_ ( cls , lowercase , lowercase="speaker_embeddings_path.json" , **lowercase ): if speaker_embeddings_dict_path is not None: _lowerCamelCase : Optional[Any] = get_file_from_repo( lowercase , lowercase , subfolder=kwargs.pop('subfolder' , lowercase ) , cache_dir=kwargs.pop('cache_dir' , lowercase ) , force_download=kwargs.pop('force_download' , lowercase ) , proxies=kwargs.pop('proxies' , lowercase ) , resume_download=kwargs.pop('resume_download' , lowercase ) , local_files_only=kwargs.pop('local_files_only' , lowercase ) , use_auth_token=kwargs.pop('use_auth_token' , lowercase ) , revision=kwargs.pop('revision' , lowercase ) , ) if speaker_embeddings_path is None: logger.warning( F'''`{os.path.join(lowercase , lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _lowerCamelCase : List[Any] = None else: with open(lowercase ) as speaker_embeddings_json: _lowerCamelCase : Union[str, Any] = json.load(lowercase ) else: _lowerCamelCase : Tuple = None _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , **lowercase ) return cls(tokenizer=lowercase , speaker_embeddings=lowercase ) def A_ ( self , lowercase , lowercase="speaker_embeddings_path.json" , lowercase="speaker_embeddings" , lowercase = False , **lowercase , ): if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowercase , lowercase , 'v2' ) , exist_ok=lowercase ) _lowerCamelCase : int = {} _lowerCamelCase : List[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _lowerCamelCase : Optional[Any] = self._load_voice_preset(lowercase ) _lowerCamelCase : Any = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , lowercase , F'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=lowercase , ) _lowerCamelCase : List[str] = os.path.join(lowercase , F'''{prompt_key}_{key}.npy''' ) _lowerCamelCase : Optional[Any] = tmp_dict with open(os.path.join(lowercase , lowercase ) , 'w' ) as fp: json.dump(lowercase , lowercase ) super().save_pretrained(lowercase , lowercase , **lowercase ) def A_ ( self , lowercase = None , **lowercase ): _lowerCamelCase : Tuple = self.speaker_embeddings[voice_preset] _lowerCamelCase : Any = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _lowerCamelCase : Union[str, Any] = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , lowercase ) , cache_dir=kwargs.pop('cache_dir' , lowercase ) , force_download=kwargs.pop('force_download' , lowercase ) , proxies=kwargs.pop('proxies' , lowercase ) , resume_download=kwargs.pop('resume_download' , lowercase ) , local_files_only=kwargs.pop('local_files_only' , lowercase ) , use_auth_token=kwargs.pop('use_auth_token' , lowercase ) , revision=kwargs.pop('revision' , lowercase ) , ) if path is None: raise ValueError( F'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _lowerCamelCase : List[str] = np.load(lowercase ) return voice_preset_dict def A_ ( self , lowercase = None ): for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , lowercase=None , lowercase=None , lowercase="pt" , lowercase=256 , lowercase=False , lowercase=True , lowercase=False , **lowercase , ): if voice_preset is not None and not isinstance(lowercase , lowercase ): if ( isinstance(lowercase , lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _lowerCamelCase : Any = self._load_voice_preset(lowercase ) else: if isinstance(lowercase , lowercase ) and not voice_preset.endswith('.npz' ): _lowerCamelCase : Optional[Any] = voice_preset + '.npz' _lowerCamelCase : Union[str, Any] = np.load(lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(lowercase , **lowercase ) _lowerCamelCase : Tuple = BatchFeature(data=lowercase , tensor_type=lowercase ) _lowerCamelCase : Any = self.tokenizer( lowercase , return_tensors=lowercase , padding='max_length' , max_length=lowercase , return_attention_mask=lowercase , return_token_type_ids=lowercase , add_special_tokens=lowercase , **lowercase , ) if voice_preset is not None: _lowerCamelCase : Optional[int] = voice_preset return encoded_text
12
0
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCamelCase__( __lowerCamelCase): def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = SMALL_MODEL_IDENTIFIER __lowerCamelCase = """pt""" __lowerCamelCase = """tf""" def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = TFAutoModel.from_pretrained(self.test_model , from_pt=UpperCamelCase_ ) model_tf.save_pretrained(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = """mock_framework""" # Framework provided - return whatever the user provides __lowerCamelCase = FeaturesManager.determine_framework(self.test_model , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(UpperCamelCase_ ) __lowerCamelCase = FeaturesManager.determine_framework(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(UpperCamelCase_ ) __lowerCamelCase = FeaturesManager.determine_framework(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: int ): # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(UpperCamelCase_ ) __lowerCamelCase = FeaturesManager.determine_framework(UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(UpperCamelCase_ ) __lowerCamelCase = FeaturesManager.determine_framework(UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = FeaturesManager.determine_framework(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) with patch("""transformers.onnx.features.is_tf_available""" , UpperCamelCase_ ): __lowerCamelCase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(UpperCamelCase_ , self.framework_pt ) # PyTorch not in environment -> use TensorFlow __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) with patch("""transformers.onnx.features.is_torch_available""" , UpperCamelCase_ ): __lowerCamelCase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(UpperCamelCase_ , self.framework_tf ) # Both in environment -> use PyTorch __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) with patch("""transformers.onnx.features.is_tf_available""" , UpperCamelCase_ ), patch( """transformers.onnx.features.is_torch_available""" , UpperCamelCase_ ): __lowerCamelCase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(UpperCamelCase_ , self.framework_pt ) # Both not in environment -> raise error __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) __lowerCamelCase = MagicMock(return_value=UpperCamelCase_ ) with patch("""transformers.onnx.features.is_tf_available""" , UpperCamelCase_ ), patch( """transformers.onnx.features.is_torch_available""" , UpperCamelCase_ ): with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = FeaturesManager.determine_framework(self.test_model )
12
from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """nielsr/canine-s""": 2048, } # Unicode defines 1,114,112 total “codepoints” _UpperCAmelCase : Tuple = 111_4112 # Below: Constants defining canonical codepoints for special, pseudo-characters. # Copied from https://github.com/google-research/language/blob/master/language/canine/special_codepoints.py _UpperCAmelCase : List[str] = 0 _UpperCAmelCase : Any = 0xE000 _UpperCAmelCase : Dict = 0xE001 _UpperCAmelCase : Optional[int] = 0xE002 _UpperCAmelCase : Tuple = 0xE003 _UpperCAmelCase : Tuple = 0xE004 # Maps special codepoints to human-readable names. _UpperCAmelCase : Dict[int, str] = { # Special symbols are represented using codepoints values that are valid, # but designated as "Private Use", meaning that they will never be assigned # characters by the Unicode Consortium, and are thus safe for use here. # # NOTE: Do *NOT* add any sort of [UNK_CHAR] here. They are explicitly # excluded and should fail with a hard error. CLS: "[CLS]", SEP: "[SEP]", BOS: "[BOS]", MASK: "[MASK]", PAD: "[PAD]", RESERVED: "[RESERVED]", } # Maps special codepoint human-readable names to their codepoint values. _UpperCAmelCase : Dict[str, int] = {name: codepoint for codepoint, name in SPECIAL_CODEPOINTS.items()} class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , snake_case=chr(snake_case ) , snake_case=chr(snake_case ) , snake_case=chr(snake_case ) , snake_case=chr(snake_case ) , snake_case=chr(snake_case ) , snake_case=chr(snake_case ) , snake_case=False , snake_case=2048 , **snake_case , ): snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else bos_token snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else eos_token snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else sep_token snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else cls_token snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else pad_token # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token super().__init__( bos_token=snake_case , eos_token=snake_case , sep_token=snake_case , cls_token=snake_case , pad_token=snake_case , mask_token=snake_case , add_prefix_space=snake_case , model_max_length=snake_case , **snake_case , ) # Creates a mapping for looking up the IDs of special symbols. snake_case_ = {} for codepoint, name in SPECIAL_CODEPOINTS.items(): snake_case_ = codepoint # Creates a mapping for looking up the string forms of special symbol IDs. snake_case_ = { codepoint: name for name, codepoint in self._special_codepoints.items() } snake_case_ = UNICODE_VOCAB_SIZE snake_case_ = len(self._special_codepoints ) @property def a ( self ): return self._unicode_vocab_size def a ( self , snake_case ): return list(snake_case ) def a ( self , snake_case ): try: return ord(snake_case ) except TypeError: raise ValueError(F'''invalid token: \'{token}\'''' ) def a ( self , snake_case ): try: if index in SPECIAL_CODEPOINTS: return SPECIAL_CODEPOINTS[index] return chr(snake_case ) except TypeError: raise ValueError(F'''invalid id: {index}''' ) def a ( self , snake_case ): return "".join(snake_case ) def a ( self , snake_case , snake_case = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] snake_case_ = cls + token_ids_a + sep if token_ids_a is not None: result += token_ids_a + sep return result def a ( self , snake_case , snake_case = None , snake_case = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) snake_case_ = [1] + ([0] * len(snake_case )) + [1] if token_ids_a is not None: result += ([0] * len(snake_case )) + [1] return result def a ( self , snake_case , snake_case = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] snake_case_ = len(cls + token_ids_a + sep ) * [0] if token_ids_a is not None: result += len(token_ids_a + sep ) * [1] return result def a ( self , snake_case , snake_case = None ): return ()
285
0
"""simple docstring""" from __future__ import annotations import os from typing import Any import requests _a : int= "https://api.github.com" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _a : Dict= BASE_URL + "/user" # https://github.com/settings/tokens _a : Union[str, Any]= os.environ.get("USER_TOKEN", "") def __UpperCAmelCase ( UpperCAmelCase_ : str ) -> dict[Any, Any]: '''simple docstring''' __snake_case : Tuple = { 'Authorization': F"token {auth_token}", 'Accept': 'application/vnd.github.v3+json', } return requests.get(UpperCAmelCase_ , headers=UpperCAmelCase_ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(f'''{key}: {value}''') else: raise ValueError("'USER_TOKEN' field cannot be empty.")
95
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : Union[str, Any]) -> Optional[int]: __snake_case : Optional[Any] = 0 def _lowercase (self : Tuple) -> int: __snake_case : Optional[Any] = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32') self.assertIsInstance(_A , _A) def _lowercase (self : str) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : List[str] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : Any) -> Optional[int]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Any = Path(_A) / 'preprocessor_config.json' __snake_case : List[Any] = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : str = CLIPConfig() # Create a dummy config file with image_proceesor_type __snake_case : List[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __snake_case : List[str] = AutoImageProcessor.from_pretrained(_A).to_dict() config_dict.pop('image_processor_type') __snake_case : Optional[int] = CLIPImageProcessor(**_A) # save in new folder model_config.save_pretrained(_A) config.save_pretrained(_A) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A) # make sure private variable is not incorrectly saved __snake_case : int = json.loads(config.to_json_string()) self.assertTrue('_processor_class' not in dict_as_saved) self.assertIsInstance(_A , _A) def _lowercase (self : Union[str, Any]) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = Path(_A) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) __snake_case : List[str] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : Optional[int]) -> Dict: with self.assertRaisesRegex( _A , 'clip-base is not a local folder and is not a valid model identifier'): __snake_case : Tuple = AutoImageProcessor.from_pretrained('clip-base') def _lowercase (self : str) -> int: with self.assertRaisesRegex( _A , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'): __snake_case : str = AutoImageProcessor.from_pretrained(_A , revision='aaaaaa') def _lowercase (self : List[Any]) -> str: with self.assertRaisesRegex( _A , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __snake_case : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model') def _lowercase (self : Optional[int]) -> List[str]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_A): __snake_case : Any = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') # If remote code is disabled, we can't load this config. with self.assertRaises(_A): __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A , trust_remote_code=_A) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor') def _lowercase (self : int) -> Optional[int]: try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A): AutoImageProcessor.register(_A , _A) with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Tuple = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = CustomImageProcessor.from_pretrained(_A) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowercase (self : List[Any]) -> Tuple: class UpperCamelCase ( lowercase ): UpperCAmelCase : str = True try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # If remote code is not set, the default is to use local __snake_case : Tuple = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote code is disabled, we load the local one. __snake_case : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote is enabled, we load from the Hub __snake_case : List[Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(not hasattr(_A , 'is_local')) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
95
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/resolve/main/config.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/config.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/config.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json', } class a ( __UpperCAmelCase ): """simple docstring""" lowerCamelCase :Dict = '''bloom''' lowerCamelCase :Optional[Any] = ['''past_key_values'''] lowerCamelCase :Any = { '''num_hidden_layers''': '''n_layer''', '''num_attention_heads''': '''n_head''', } def __init__( self , lowerCAmelCase_=25_08_80 , lowerCAmelCase_=64 , lowerCAmelCase_=2 , lowerCAmelCase_=8 , lowerCAmelCase_=1E-5 , lowerCAmelCase_=0.02 , lowerCAmelCase_=True , lowerCAmelCase_=1 , lowerCAmelCase_=2 , lowerCAmelCase_=False , lowerCAmelCase_=0.0 , lowerCAmelCase_=0.0 , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ , ) -> Union[str, Any]: _A = vocab_size # Backward compatibility with n_embed kwarg _A = kwargs.pop("""n_embed""" , UpperCamelCase__ ) _A = hidden_size if n_embed is None else n_embed _A = n_layer _A = n_head _A = layer_norm_epsilon _A = initializer_range _A = use_cache _A = pretraining_tp _A = apply_residual_connection_post_layernorm _A = hidden_dropout _A = attention_dropout _A = bos_token_id _A = eos_token_id _A = slow_but_exact super().__init__(bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) class a ( __UpperCAmelCase ): """simple docstring""" lowerCamelCase :List[str] = version.parse('''1.12''' ) def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ = "default" , lowerCAmelCase_ = None , lowerCAmelCase_ = False , ) -> int: super().__init__(UpperCamelCase__ , task=UpperCamelCase__ , patching_specs=UpperCamelCase__ , use_past=UpperCamelCase__ ) if not getattr(self._config , """pad_token_id""" , UpperCamelCase__ ): # TODO: how to do that better? _A = 0 @property def UpperCAmelCase ( self ) -> List[str]: _A = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(UpperCamelCase__ , direction="""inputs""" , inverted_values_shape=UpperCamelCase__ ) _A = {0: """batch""", 1: """past_sequence + sequence"""} else: _A = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCAmelCase ( self ) -> List[str]: return self._config.n_layer @property def UpperCAmelCase ( self ) -> Any: return self._config.n_head @property def UpperCAmelCase ( self ) -> List[str]: return 1E-3 def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = -1 , lowerCAmelCase_ = -1 , lowerCAmelCase_ = False , lowerCAmelCase_ = None , ) -> int: _A = super(UpperCamelCase__ , self ).generate_dummy_inputs( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) # We need to order the input in the way they appears in the forward() _A = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch _A , _A = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values _A = seqlen + 2 _A = self._config.hidden_size // self.num_attention_heads _A = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) _A = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) _A = [ (torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) for _ in range(self.num_layers ) ] _A = common_inputs["""attention_mask"""] if self.use_past: _A = ordered_inputs["""attention_mask"""].dtype _A = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(UpperCamelCase__ , UpperCamelCase__ , dtype=UpperCamelCase__ )] , dim=1 ) return ordered_inputs @property def UpperCAmelCase ( self ) -> str: return 13
180
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor _A = logging.get_logger(__name__) class A ( __UpperCAmelCase ): def __init__( self, *UpperCamelCase__, **UpperCamelCase__ ): """simple docstring""" warnings.warn( '''The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use BeitImageProcessor instead.''', UpperCamelCase__, ) super().__init__(*UpperCamelCase__, **UpperCamelCase__ )
278
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __SCREAMING_SNAKE_CASE :Tuple = {'''configuration_glpn''': ['''GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GLPNConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE :Tuple = ['''GLPNFeatureExtractor'''] __SCREAMING_SNAKE_CASE :Tuple = ['''GLPNImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE :Optional[Any] = [ '''GLPN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GLPNForDepthEstimation''', '''GLPNLayer''', '''GLPNModel''', '''GLPNPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE :str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
371
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __SCREAMING_SNAKE_CASE :Tuple = '''\ ''' __SCREAMING_SNAKE_CASE :Union[str, Any] = ''' Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity ''' __SCREAMING_SNAKE_CASE :List[Any] = ''' Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to \'cuda\' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"] >>> results = perplexity.compute(model_id=\'gpt2\', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 78.22 >>> print(round(results["perplexities"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = datasets.load_dataset("wikitext", ... "wikitext-2-raw-v1", ... split="test")["text"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=\'\'] >>> results = perplexity.compute(model_id=\'gpt2\', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 60.35 >>> print(round(results["perplexities"][0], 2)) 81.12 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ ( datasets.Metric ): def lowercase ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "input_texts": datasets.Value("string" ), } ) , reference_urls=["https://huggingface.co/docs/transformers/perplexity"] , ) def lowercase ( self : List[Any] , snake_case_ : Optional[Any] , snake_case_ : List[Any] , snake_case_ : int = 1_6 , snake_case_ : bool = True , snake_case_ : int=None ): if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": _UpperCAmelCase = "cuda" else: _UpperCAmelCase = "cuda" if torch.cuda.is_available() else "cpu" _UpperCAmelCase = AutoModelForCausalLM.from_pretrained(snake_case_ ) _UpperCAmelCase = model.to(snake_case_ ) _UpperCAmelCase = AutoTokenizer.from_pretrained(snake_case_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: _UpperCAmelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(snake_case_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"pad_token": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" _UpperCAmelCase = model.config.max_length - 1 else: _UpperCAmelCase = model.config.max_length _UpperCAmelCase = tokenizer( snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , return_tensors="pt" , return_attention_mask=snake_case_ , ).to(snake_case_ ) _UpperCAmelCase = encodings["input_ids"] _UpperCAmelCase = encodings["attention_mask"] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." _UpperCAmelCase = [] _UpperCAmelCase = CrossEntropyLoss(reduction="none" ) for start_index in logging.tqdm(range(0 , len(snake_case_ ) , snake_case_ ) ): _UpperCAmelCase = min(start_index + batch_size , len(snake_case_ ) ) _UpperCAmelCase = encoded_texts[start_index:end_index] _UpperCAmelCase = attn_masks[start_index:end_index] if add_start_token: _UpperCAmelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(snake_case_ ) _UpperCAmelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) _UpperCAmelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(snake_case_ ), attn_mask] , dim=1 ) _UpperCAmelCase = encoded_batch with torch.no_grad(): _UpperCAmelCase = model(snake_case_ , attention_mask=snake_case_ ).logits _UpperCAmelCase = out_logits[..., :-1, :].contiguous() _UpperCAmelCase = labels[..., 1:].contiguous() _UpperCAmelCase = attn_mask[..., 1:].contiguous() _UpperCAmelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , snake_case_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(snake_case_ )}
156
0
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE :List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Tuple = [ ['''attention''', '''attn'''], ['''encoder_attention''', '''encoder_attn'''], ['''q_lin''', '''q_proj'''], ['''k_lin''', '''k_proj'''], ['''v_lin''', '''v_proj'''], ['''out_lin''', '''out_proj'''], ['''norm_embeddings''', '''layernorm_embedding'''], ['''position_embeddings''', '''embed_positions'''], ['''embeddings''', '''embed_tokens'''], ['''ffn.lin''', '''fc'''], ] def _lowerCAmelCase ( lowerCAmelCase_ :Optional[int] )->List[Any]: '''simple docstring''' if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: snake_case_ = k.replace(lowerCAmelCase_ , lowerCAmelCase_ ) if k.startswith("encoder" ): snake_case_ = k.replace(".attn" , ".self_attn" ) snake_case_ = k.replace("norm1" , "self_attn_layer_norm" ) snake_case_ = k.replace("norm2" , "final_layer_norm" ) elif k.startswith("decoder" ): snake_case_ = k.replace("norm1" , "self_attn_layer_norm" ) snake_case_ = k.replace("norm2" , "encoder_attn_layer_norm" ) snake_case_ = k.replace("norm3" , "final_layer_norm" ) return k def _lowerCAmelCase ( lowerCAmelCase_ :Any )->Any: '''simple docstring''' snake_case_ = [ "model.encoder.layernorm_embedding.weight", "model.encoder.layernorm_embedding.bias", "model.decoder.layernorm_embedding.weight", "model.decoder.layernorm_embedding.bias", ] for k in keys: snake_case_ = sd.pop(lowerCAmelCase_ ) snake_case_ = k.replace("layernorm_embedding" , "layer_norm" ) assert new_k not in sd snake_case_ = v SCREAMING_SNAKE_CASE :Tuple = ['''START'''] @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase_ :Tuple , lowerCAmelCase_ :List[str] , lowerCAmelCase_ :Union[str, Any] )->List[Any]: '''simple docstring''' snake_case_ = torch.load(lowerCAmelCase_ , map_location="cpu" ) snake_case_ = model["model"] snake_case_ = BlenderbotConfig.from_json_file(lowerCAmelCase_ ) snake_case_ = BlenderbotForConditionalGeneration(lowerCAmelCase_ ) snake_case_ = m.model.state_dict().keys() snake_case_ = [] snake_case_ = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue snake_case_ = rename_state_dict_key(lowerCAmelCase_ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: snake_case_ = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(lowerCAmelCase_ ) m.model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) m.half() m.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE :Any = argparse.ArgumentParser() # Required parameters parser.add_argument('''--src_path''', type=str, help='''like blenderbot-model.bin''') parser.add_argument('''--save_dir''', default='''hf_blenderbot''', type=str, help='''Where to save converted model.''') parser.add_argument( '''--hf_config_json''', default='''blenderbot-3b-config.json''', type=str, help='''Path to config to use''' ) SCREAMING_SNAKE_CASE :Optional[Any] = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
159
import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE :Optional[Any] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase_ :List[Any] )->int: '''simple docstring''' print("Loading config file..." ) def flatten_yaml_as_dict(lowerCAmelCase_ :List[Any] , lowerCAmelCase_ :Optional[int]="" , lowerCAmelCase_ :int="." ): snake_case_ = [] for k, v in d.items(): snake_case_ = parent_key + sep + k if parent_key else k if isinstance(lowerCAmelCase_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(lowerCAmelCase_ , lowerCAmelCase_ , sep=lowerCAmelCase_ ).items() ) else: items.append((new_key, v) ) return dict(lowerCAmelCase_ ) snake_case_ = argparse.Namespace() with open(lowerCAmelCase_ , "r" ) as yaml_file: try: snake_case_ = yaml.load(lowerCAmelCase_ , Loader=yaml.FullLoader ) snake_case_ = flatten_yaml_as_dict(lowerCAmelCase_ ) for k, v in flat_cfg.items(): setattr(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) except yaml.YAMLError as exc: logger.error("Error while loading config file: {}. Error message: {}".format(lowerCAmelCase_ , str(lowerCAmelCase_ ) ) ) return config def _lowerCAmelCase ( lowerCAmelCase_ :List[str] , lowerCAmelCase_ :Tuple )->Union[str, Any]: '''simple docstring''' snake_case_ = MobileViTVaConfig() snake_case_ = False # dataset if task_name.startswith("imagenet1k_" ): snake_case_ = 1_000 if int(task_name.strip().split("_" )[-1] ) == 384: snake_case_ = 384 else: snake_case_ = 256 snake_case_ = "imagenet-1k-id2label.json" elif task_name.startswith("imagenet21k_to_1k_" ): snake_case_ = 21_000 if int(task_name.strip().split("_" )[-1] ) == 384: snake_case_ = 384 else: snake_case_ = 256 snake_case_ = "imagenet-22k-id2label.json" elif task_name.startswith("ade20k_" ): snake_case_ = 151 snake_case_ = 512 snake_case_ = "ade20k-id2label.json" snake_case_ = True elif task_name.startswith("voc_" ): snake_case_ = 21 snake_case_ = 512 snake_case_ = "pascal-voc-id2label.json" snake_case_ = True # orig_config snake_case_ = load_orig_config_file(lowerCAmelCase_ ) assert getattr(lowerCAmelCase_ , "model.classification.name" , -1 ) == "mobilevit_v2", "Invalid model" snake_case_ = getattr(lowerCAmelCase_ , "model.classification.mitv2.width_multiplier" , 1.0 ) assert ( getattr(lowerCAmelCase_ , "model.classification.mitv2.attn_norm_layer" , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" snake_case_ = getattr(lowerCAmelCase_ , "model.classification.activation.name" , "swish" ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: snake_case_ = getattr(lowerCAmelCase_ , "model.segmentation.output_stride" , 16 ) if "_deeplabv3" in task_name: snake_case_ = getattr(lowerCAmelCase_ , "model.segmentation.deeplabv3.aspp_rates" , [12, 24, 36] ) snake_case_ = getattr(lowerCAmelCase_ , "model.segmentation.deeplabv3.aspp_out_channels" , 512 ) snake_case_ = getattr(lowerCAmelCase_ , "model.segmentation.deeplabv3.aspp_dropout" , 0.1 ) # id2label snake_case_ = "huggingface/label-files" snake_case_ = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type="dataset" ) , "r" ) ) snake_case_ = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def _lowerCAmelCase ( lowerCAmelCase_ :Any , lowerCAmelCase_ :Optional[Any] , lowerCAmelCase_ :Optional[Any] )->Optional[Any]: '''simple docstring''' snake_case_ = dct.pop(lowerCAmelCase_ ) snake_case_ = val def _lowerCAmelCase ( lowerCAmelCase_ :Dict , lowerCAmelCase_ :int=False )->Dict: '''simple docstring''' if base_model: snake_case_ = "" else: snake_case_ = "mobilevitv2." snake_case_ = [] for k in state_dict.keys(): if k[:8] == "encoder.": snake_case_ = k[8:] else: snake_case_ = k if ".block." in k: snake_case_ = k_new.replace(".block." , "." ) if ".conv." in k: snake_case_ = k_new.replace(".conv." , ".convolution." ) if ".norm." in k: snake_case_ = k_new.replace(".norm." , ".normalization." ) if "conv_1." in k: snake_case_ = k_new.replace("conv_1." , F'''{model_prefix}conv_stem.''' ) for i in [1, 2]: if F'''layer_{i}.''' in k: snake_case_ = k_new.replace(F'''layer_{i}.''' , F'''{model_prefix}encoder.layer.{i-1}.layer.''' ) if ".exp_1x1." in k: snake_case_ = k_new.replace(".exp_1x1." , ".expand_1x1." ) if ".red_1x1." in k: snake_case_ = k_new.replace(".red_1x1." , ".reduce_1x1." ) for i in [3, 4, 5]: if F'''layer_{i}.0.''' in k: snake_case_ = k_new.replace(F'''layer_{i}.0.''' , F'''{model_prefix}encoder.layer.{i-1}.downsampling_layer.''' ) if F'''layer_{i}.1.local_rep.0.''' in k: snake_case_ = k_new.replace(F'''layer_{i}.1.local_rep.0.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_kxk.''' ) if F'''layer_{i}.1.local_rep.1.''' in k: snake_case_ = k_new.replace(F'''layer_{i}.1.local_rep.1.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_1x1.''' ) for i in [3, 4, 5]: if i == 3: snake_case_ = [0, 1] elif i == 4: snake_case_ = [0, 1, 2, 3] elif i == 5: snake_case_ = [0, 1, 2] for j in j_in: if F'''layer_{i}.1.global_rep.{j}.''' in k: snake_case_ = k_new.replace( F'''layer_{i}.1.global_rep.{j}.''' , F'''{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.''' ) if F'''layer_{i}.1.global_rep.{j+1}.''' in k: snake_case_ = k_new.replace( F'''layer_{i}.1.global_rep.{j+1}.''' , F'''{model_prefix}encoder.layer.{i-1}.layernorm.''' ) if F'''layer_{i}.1.conv_proj.''' in k: snake_case_ = k_new.replace(F'''layer_{i}.1.conv_proj.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_projection.''' ) if "pre_norm_attn.0." in k: snake_case_ = k_new.replace("pre_norm_attn.0." , "layernorm_before." ) if "pre_norm_attn.1." in k: snake_case_ = k_new.replace("pre_norm_attn.1." , "attention." ) if "pre_norm_ffn.0." in k: snake_case_ = k_new.replace("pre_norm_ffn.0." , "layernorm_after." ) if "pre_norm_ffn.1." in k: snake_case_ = k_new.replace("pre_norm_ffn.1." , "ffn.conv1." ) if "pre_norm_ffn.3." in k: snake_case_ = k_new.replace("pre_norm_ffn.3." , "ffn.conv2." ) if "classifier.1." in k: snake_case_ = k_new.replace("classifier.1." , "classifier." ) if "seg_head." in k: snake_case_ = k_new.replace("seg_head." , "segmentation_head." ) if ".aspp_layer." in k: snake_case_ = k_new.replace(".aspp_layer." , "." ) if ".aspp_pool." in k: snake_case_ = k_new.replace(".aspp_pool." , "." ) rename_keys.append((k, k_new) ) return rename_keys def _lowerCAmelCase ( lowerCAmelCase_ :Optional[Any] )->Optional[int]: '''simple docstring''' snake_case_ = [] for k in state_dict.keys(): if k.startswith("seg_head.aux_head." ): keys_to_ignore.append(lowerCAmelCase_ ) for k in keys_to_ignore: state_dict.pop(lowerCAmelCase_ , lowerCAmelCase_ ) def _lowerCAmelCase ( )->List[Any]: '''simple docstring''' snake_case_ = "http://images.cocodataset.org/val2017/000000039769.jpg" # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" snake_case_ = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase_ :Optional[int] , lowerCAmelCase_ :Dict , lowerCAmelCase_ :Optional[int] , lowerCAmelCase_ :Dict )->Dict: '''simple docstring''' snake_case_ = get_mobilevitva_config(lowerCAmelCase_ , lowerCAmelCase_ ) # load original state_dict snake_case_ = torch.load(lowerCAmelCase_ , map_location="cpu" ) # load huggingface model if task_name.startswith("ade20k_" ) or task_name.startswith("voc_" ): snake_case_ = MobileViTVaForSemanticSegmentation(lowerCAmelCase_ ).eval() snake_case_ = False else: snake_case_ = MobileViTVaForImageClassification(lowerCAmelCase_ ).eval() snake_case_ = False # remove and rename some keys of load the original model snake_case_ = checkpoint remove_unused_keys(lowerCAmelCase_ ) snake_case_ = create_rename_keys(lowerCAmelCase_ , base_model=lowerCAmelCase_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # load modified state_dict model.load_state_dict(lowerCAmelCase_ ) # Check outputs on an image, prepared by MobileViTImageProcessor snake_case_ = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) snake_case_ = image_processor(images=prepare_img() , return_tensors="pt" ) snake_case_ = model(**lowerCAmelCase_ ) # verify classification model if task_name.startswith("imagenet" ): snake_case_ = outputs.logits snake_case_ = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) if task_name.startswith("imagenet1k_256" ) and config.width_multiplier == 1.0: # expected_logits for base variant snake_case_ = torch.tensor([-1.63_36e00, -7.32_04e-02, -5.18_83e-01] ) assert torch.allclose(logits[0, :3] , lowerCAmelCase_ , atol=1e-4 ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(F'''Saving model {task_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase_ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''imagenet1k_256''', type=str, help=( '''Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . ''' ''' Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 ''' ), choices=[ '''imagenet1k_256''', '''imagenet1k_384''', '''imagenet21k_to_1k_256''', '''imagenet21k_to_1k_384''', '''ade20k_deeplabv3''', '''voc_deeplabv3''', ], ) parser.add_argument( '''--orig_checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument('''--orig_config_path''', required=True, type=str, help='''Path to the original config file.''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) SCREAMING_SNAKE_CASE :int = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
159
1
'''simple docstring''' lowerCAmelCase : Union[str, Any] =[ '''VerificationMode''', '''Version''', '''disable_progress_bar''', '''enable_progress_bar''', '''is_progress_bar_enabled''', '''experimental''', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
366
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : list ): if len(__lowerCamelCase ) <= 1: return lst lowercase_ :Optional[Any] = 1 while i < len(__lowerCamelCase ): if lst[i - 1] <= lst[i]: i += 1 else: lowercase_ , lowercase_ :int = lst[i], lst[i - 1] i -= 1 if i == 0: lowercase_ :Dict = 1 return lst if __name__ == "__main__": lowerCAmelCase : Any =input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase : List[str] =[int(item) for item in user_input.split(''',''')] print(gnome_sort(unsorted))
147
0
"""simple docstring""" from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class snake_case ( __snake_case ): def __init__( self : Optional[int] , UpperCamelCase__ : Optional[NestedDataStructureLike[PathLike]] = None , UpperCamelCase__ : Optional[NamedSplit] = None , UpperCamelCase__ : Optional[Features] = None , UpperCamelCase__ : str = None , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , UpperCamelCase__ : Optional[int] = None , **UpperCamelCase__ : int , )-> Dict: '''simple docstring''' __lowerCAmelCase: Tuple = path_or_paths __lowerCAmelCase: List[str] = split if split or isinstance(UpperCamelCase__ , UpperCamelCase__) else "train" __lowerCAmelCase: Union[str, Any] = features __lowerCAmelCase: Dict = cache_dir __lowerCAmelCase: Any = keep_in_memory __lowerCAmelCase: str = streaming __lowerCAmelCase: Optional[Any] = num_proc __lowerCAmelCase: str = kwargs @abstractmethod def lowercase_ ( self : int)-> Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]: '''simple docstring''' pass class snake_case ( __snake_case ): def __init__( self : int , UpperCamelCase__ : Optional[Features] = None , UpperCamelCase__ : str = None , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , UpperCamelCase__ : Optional[int] = None , **UpperCamelCase__ : List[str] , )-> List[Any]: '''simple docstring''' __lowerCAmelCase: List[Any] = features __lowerCAmelCase: Tuple = cache_dir __lowerCAmelCase: List[str] = keep_in_memory __lowerCAmelCase: Optional[Any] = streaming __lowerCAmelCase: List[Any] = num_proc __lowerCAmelCase: str = kwargs @abstractmethod def lowercase_ ( self : Tuple)-> Union[Dataset, IterableDataset]: '''simple docstring''' pass
217
"""simple docstring""" def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: Optional[Any] = 1 for i in range(1 , num + 1 ): fact *= i return fact def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: List[str] = 0 while number > 0: __lowerCAmelCase: Any = number % 1_0 sum_of_digits += last_digit __lowerCAmelCase: List[Any] = number // 1_0 # Removing the last_digit from the given number return sum_of_digits def a__ ( __SCREAMING_SNAKE_CASE = 1_0_0 ) -> int: __lowerCAmelCase: Tuple = factorial(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Optional[int] = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
217
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE_ = { 'configuration_luke': ['LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LukeConfig'], 'tokenization_luke': ['LukeTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ 'LUKE_PRETRAINED_MODEL_ARCHIVE_LIST', 'LukeForEntityClassification', 'LukeForEntityPairClassification', 'LukeForEntitySpanClassification', 'LukeForMultipleChoice', 'LukeForQuestionAnswering', 'LukeForSequenceClassification', 'LukeForTokenClassification', 'LukeForMaskedLM', 'LukeModel', 'LukePreTrainedModel', ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
189
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: str , lowerCAmelCase: float | Decimal , lowerCAmelCase: float = 10**-10 ) -> float: _UpperCAmelCase : Optional[int] = a while True: _UpperCAmelCase : Tuple = Decimal(lowerCAmelCase ) - ( Decimal(eval(lowerCAmelCase ) ) / Decimal(eval(str(diff(lowerCAmelCase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase ) ) < precision: # noqa: S307 return float(lowerCAmelCase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'''The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}''') # Find root of polynomial print(F'''The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}''') # Find Square Root of 5 print(F'''The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}''') # Exponential Roots print(F'''The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}''')
189
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() __snake_case =logging.get_logger(__name__) def a_ ( lowerCamelCase : Union[str, Any] , lowerCamelCase : List[str]=False ): lowerCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'deit.embeddings.cls_token'), ('dist_token', 'deit.embeddings.distillation_token'), ('patch_embed.proj.weight', 'deit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'deit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'deit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" lowerCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('deit' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('norm.weight', 'deit.layernorm.weight'), ('norm.bias', 'deit.layernorm.bias'), ('head.weight', 'cls_classifier.weight'), ('head.bias', 'cls_classifier.bias'), ('head_dist.weight', 'distillation_classifier.weight'), ('head_dist.bias', 'distillation_classifier.bias'), ] ) return rename_keys def a_ ( lowerCamelCase : Tuple , lowerCamelCase : Union[str, Any] , lowerCamelCase : Union[str, Any]=False ): for i in range(config.num_hidden_layers ): if base_model: lowerCAmelCase = '' else: lowerCAmelCase = 'deit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) lowerCAmelCase = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase = in_proj_weight[ : config.hidden_size, : ] lowerCAmelCase = in_proj_bias[: config.hidden_size] lowerCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase = in_proj_weight[ -config.hidden_size :, : ] lowerCAmelCase = in_proj_bias[-config.hidden_size :] def a_ ( lowerCamelCase : Dict , lowerCamelCase : List[Any] , lowerCamelCase : str ): lowerCAmelCase = dct.pop(lowerCamelCase ) lowerCAmelCase = val def a_ ( ): lowerCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCAmelCase = Image.open(requests.get(lowerCamelCase , stream=lowerCamelCase ).raw ) return im @torch.no_grad() def a_ ( lowerCamelCase : List[str] , lowerCamelCase : Dict ): lowerCAmelCase = DeiTConfig() # all deit models have fine-tuned heads lowerCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size lowerCAmelCase = 1000 lowerCAmelCase = 'huggingface/label-files' lowerCAmelCase = 'imagenet-1k-id2label.json' lowerCAmelCase = json.load(open(hf_hub_download(lowerCamelCase , lowerCamelCase , repo_type='dataset' ) , 'r' ) ) lowerCAmelCase = {int(lowerCamelCase ): v for k, v in idalabel.items()} lowerCAmelCase = idalabel lowerCAmelCase = {v: k for k, v in idalabel.items()} lowerCAmelCase = int(deit_name[-6:-4] ) lowerCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('tiny' ): lowerCAmelCase = 192 lowerCAmelCase = 768 lowerCAmelCase = 12 lowerCAmelCase = 3 elif deit_name[9:].startswith('small' ): lowerCAmelCase = 384 lowerCAmelCase = 1536 lowerCAmelCase = 12 lowerCAmelCase = 6 if deit_name[9:].startswith('base' ): pass elif deit_name[4:].startswith('large' ): lowerCAmelCase = 1024 lowerCAmelCase = 4096 lowerCAmelCase = 24 lowerCAmelCase = 16 # load original model from timm lowerCAmelCase = timm.create_model(lowerCamelCase , pretrained=lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowerCAmelCase = timm_model.state_dict() lowerCAmelCase = create_rename_keys(lowerCamelCase , lowerCamelCase ) for src, dest in rename_keys: rename_key(lowerCamelCase , lowerCamelCase , lowerCamelCase ) read_in_q_k_v(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # load HuggingFace model lowerCAmelCase = DeiTForImageClassificationWithTeacher(lowerCamelCase ).eval() model.load_state_dict(lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor lowerCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 lowerCAmelCase = DeiTImageProcessor(size=lowerCamelCase , crop_size=config.image_size ) lowerCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) lowerCAmelCase = encoding['pixel_values'] lowerCAmelCase = model(lowerCamelCase ) lowerCAmelCase = timm_model(lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCamelCase , outputs.logits , atol=1e-3 ) Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) print(f'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCamelCase ) if __name__ == "__main__": __snake_case =argparse.ArgumentParser() # Required parameters parser.add_argument( """--deit_name""", default="""vit_deit_base_distilled_patch16_224""", type=str, help="""Name of the DeiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) __snake_case =parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
4
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : List[str] = DiTPipeline __UpperCAmelCase : Tuple = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS __UpperCAmelCase : Any = PipelineTesterMixin.required_optional_params - { "latents", "num_images_per_prompt", "callback", "callback_steps", } __UpperCAmelCase : Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> str: torch.manual_seed(0 ) __snake_case : Dict = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=lowerCamelCase , activation_fn="gelu-approximate" , num_embeds_ada_norm=1000 , norm_type="ada_norm_zero" , norm_elementwise_affine=lowerCamelCase , ) __snake_case : Union[str, Any] = AutoencoderKL() __snake_case : int = DDIMScheduler() __snake_case : Tuple = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def __snake_case ( self : Optional[Any] , lowerCamelCase : Any , lowerCamelCase : Tuple=0 ) -> int: if str(lowerCamelCase ).startswith("mps" ): __snake_case : str = torch.manual_seed(lowerCamelCase ) else: __snake_case : Dict = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __snake_case : Optional[Any] = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __snake_case ( self : str ) -> List[str]: __snake_case : Tuple = "cpu" __snake_case : int = self.get_dummy_components() __snake_case : str = self.pipeline_class(**lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : Tuple = self.get_dummy_inputs(lowerCamelCase ) __snake_case : Optional[Any] = pipe(**lowerCamelCase ).images __snake_case : List[str] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) __snake_case : int = np.array([0.29_46, 0.66_01, 0.43_29, 0.32_96, 0.41_44, 0.53_19, 0.72_73, 0.50_13, 0.44_57] ) __snake_case : Optional[int] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase , 1E-3 ) def __snake_case ( self : List[str] ) -> Tuple: self._test_inference_batch_single_identical(relax_max_difference=lowerCamelCase , expected_max_diff=1E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __snake_case ( self : Tuple ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @require_torch_gpu @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Optional[Any] ) -> str: super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : List[Any] ) -> Any: __snake_case : Any = torch.manual_seed(0 ) __snake_case : List[Any] = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) __snake_case : Optional[int] = ["vase", "umbrella", "white shark", "white wolf"] __snake_case : Optional[Any] = pipe.get_label_ids(lowerCamelCase ) __snake_case : List[Any] = pipe(lowerCamelCase , generator=lowerCamelCase , num_inference_steps=40 , output_type="np" ).images for word, image in zip(lowerCamelCase , lowerCamelCase ): __snake_case : int = load_numpy( F'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy' ) assert np.abs((expected_image - image).max() ) < 1E-2 def __snake_case ( self : Union[str, Any] ) -> int: __snake_case : Any = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) __snake_case : int = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) __snake_case : Tuple = ["vase", "umbrella"] __snake_case : List[str] = pipe.get_label_ids(lowerCamelCase ) __snake_case : Tuple = torch.manual_seed(0 ) __snake_case : str = pipe(lowerCamelCase , generator=lowerCamelCase , num_inference_steps=25 , output_type="np" ).images for word, image in zip(lowerCamelCase , lowerCamelCase ): __snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" F'/dit/{word}_512.npy' ) assert np.abs((expected_image - image).max() ) < 1E-1
123
0
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase_ = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowerCamelCase_ = { """facebook/blenderbot_small-90M""": 5_1_2, } class a_ ( a_ ): '''simple docstring''' __a: Optional[int] = VOCAB_FILES_NAMES __a: List[str] = PRETRAINED_VOCAB_FILES_MAP __a: Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __a: List[str] = BlenderbotSmallTokenizer def __init__( self , lowercase_=None , lowercase_=None , lowercase_="<|endoftext|>" , lowercase_="<|endoftext|>" , lowercase_="<|endoftext|>" , lowercase_=False , lowercase_=True , **lowercase_ , ) -> Any: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowercase_ , merges=lowercase_ , add_prefix_space=lowercase_ , trim_offsets=lowercase_ , ) , bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , **lowercase_ , ) lowerCAmelCase_ = add_prefix_space def _lowercase ( self , lowercase_ , lowercase_=None ) -> int: '''simple docstring''' lowerCAmelCase_ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowercase ( self , lowercase_ , lowercase_ = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ = [self.sep_token_id] lowerCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
14
def lowerCamelCase ( a_ ) -> bool: lowerCAmelCase_ = set() # To detect a back edge, keep track of vertices currently in the recursion stack lowerCAmelCase_ = set() return any( node not in visited and depth_first_search(a_ , a_ , a_ , a_ ) for node in graph ) def lowerCamelCase ( a_ , a_ , a_ , a_ ) -> bool: visited.add(a_ ) rec_stk.add(a_ ) for node in graph[vertex]: if node not in visited: if depth_first_search(a_ , a_ , a_ , a_ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(a_ ) return False if __name__ == "__main__": from doctest import testmod testmod()
14
1
"""simple docstring""" import math def _snake_case ( _snake_case : int ) -> list: '''simple docstring''' _A = [True] * n _A = False _A = False _A = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): _A = i * 2 while index < n: _A = False _A = index + i _A = [2] for i in range(3 , _snake_case , 2 ): if is_prime[i]: primes.append(_snake_case ) return primes def _snake_case ( _snake_case : int = 99_99_66_66_33_33 ) -> int: '''simple docstring''' _A = math.floor(math.sqrt(_snake_case ) ) + 1_00 _A = prime_sieve(_snake_case ) _A = 0 _A = 0 _A = primes[prime_index] while (last_prime**2) <= limit: _A = primes[prime_index + 1] _A = last_prime**2 _A = next_prime**2 # Get numbers divisible by lps(current) _A = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) _A = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps _A = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair _A = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
315
"""simple docstring""" import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowerCAmelCase ) , '''Tatoeba directory does not exist.''' ) class lowercase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self : Optional[Any] ): _A = tempfile.mkdtemp() return TatoebaConverter(save_dir=_UpperCAmelCase ) @slow def lowerCAmelCase_ ( self : Optional[int] ): self.resolver.convert_models(['heb-eng'] ) @slow def lowerCAmelCase_ ( self : Optional[Any] ): _A , _A = self.resolver.write_model_card('opus-mt-he-en' , dry_run=_UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
315
1
from datetime import datetime import matplotlib.pyplot as plt import torch def lowerCAmelCase_ ( _snake_case : Optional[int] ) -> Union[str, Any]: '''simple docstring''' for param in module.parameters(): __magic_name__ : int = False def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = "cuda" if torch.cuda.is_available() else "cpu" if torch.backends.mps.is_available() and torch.backends.mps.is_built(): __magic_name__ : List[Any] = "mps" if device == "mps": print( "WARNING: MPS currently doesn't seem to work, and messes up backpropagation without any visible torch" " errors. I recommend using CUDA on a colab notebook or CPU instead if you're facing inexplicable issues" " with generations." ) return device def lowerCAmelCase_ ( _snake_case : Any ) -> Optional[int]: '''simple docstring''' __magic_name__ : Optional[Any] = plt.imshow(SCREAMING_SNAKE_CASE_ ) fig.axes.get_xaxis().set_visible(SCREAMING_SNAKE_CASE_ ) fig.axes.get_yaxis().set_visible(SCREAMING_SNAKE_CASE_ ) plt.show() def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' __magic_name__ : Optional[int] = datetime.now() __magic_name__ : Tuple = current_time.strftime("%H:%M:%S" ) return timestamp
363
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : Optional[Any] = logging.get_logger(__name__) snake_case : Union[str, Any] = { "transfo-xl-wt103": "https://huggingface.co/transfo-xl-wt103/resolve/main/config.json", } class _snake_case ( snake_case ): UpperCamelCase__ = 'transfo-xl' UpperCamelCase__ = ['mems'] UpperCamelCase__ = { 'n_token': 'vocab_size', 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , _a=267_735 , _a=[20_000, 40_000, 200_000] , _a=1_024 , _a=1_024 , _a=16 , _a=64 , _a=4_096 , _a=4 , _a=False , _a=18 , _a=1_600 , _a=1_000 , _a=True , _a=True , _a=0 , _a=-1 , _a=True , _a=0.1 , _a=0.0 , _a=True , _a="normal" , _a=0.01 , _a=0.01 , _a=0.02 , _a=1e-5 , _a=0 , **_a , ): __magic_name__ : List[Any] = vocab_size __magic_name__ : Dict = [] self.cutoffs.extend(_a ) if proj_share_all_but_first: __magic_name__ : List[str] = [False] + [True] * len(self.cutoffs ) else: __magic_name__ : Optional[Any] = [False] + [False] * len(self.cutoffs ) __magic_name__ : Optional[int] = d_model __magic_name__ : str = d_embed __magic_name__ : Optional[Any] = d_head __magic_name__ : Optional[int] = d_inner __magic_name__ : List[str] = div_val __magic_name__ : List[str] = pre_lnorm __magic_name__ : Union[str, Any] = n_layer __magic_name__ : Optional[int] = n_head __magic_name__ : str = mem_len __magic_name__ : int = same_length __magic_name__ : Dict = attn_type __magic_name__ : int = clamp_len __magic_name__ : Optional[int] = sample_softmax __magic_name__ : List[Any] = adaptive __magic_name__ : Optional[int] = dropout __magic_name__ : Optional[int] = dropatt __magic_name__ : Optional[Any] = untie_r __magic_name__ : List[str] = init __magic_name__ : Any = init_range __magic_name__ : Optional[int] = proj_init_std __magic_name__ : List[Any] = init_std __magic_name__ : List[Any] = layer_norm_epsilon super().__init__(eos_token_id=_a , **_a ) @property def SCREAMING_SNAKE_CASE ( self ): # Message copied from Transformer-XL documentation logger.info(f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def SCREAMING_SNAKE_CASE ( self , _a ): # Message copied from Transformer-XL documentation raise NotImplementedError( f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
41
0
'''simple docstring''' import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( _snake_case , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ "num_inference_steps", "generator", "latents", "output_type", "return_dict", "callback", "callback_steps", ] ) @property def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = UNetaDModel.from_pretrained( """diffusers/consistency-models-test""" , subfolder="""test_unet""" , ) return unet @property def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' A_ = UNetaDModel.from_pretrained( """diffusers/consistency-models-test""" , subfolder="""test_unet_class_cond""" , ) return unet def snake_case_ ( self , UpperCamelCase__=False ) -> Dict: '''simple docstring''' if class_cond: A_ = self.dummy_cond_unet else: A_ = self.dummy_uncond_unet # Default to CM multistep sampler A_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) A_ = { """unet""": unet, """scheduler""": scheduler, } return components def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__=0 ) -> List[Any]: '''simple docstring''' if str(UpperCamelCase__ ).startswith("""mps""" ): A_ = torch.manual_seed(UpperCamelCase__ ) else: A_ = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A_ = { """batch_size""": 1, """num_inference_steps""": None, """timesteps""": [22, 0], """generator""": generator, """output_type""": """np""", } return inputs def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = """cpu""" # ensure determinism for the device-dependent torch.Generator A_ = self.get_dummy_components() A_ = ConsistencyModelPipeline(**UpperCamelCase__ ) A_ = pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_dummy_inputs(UpperCamelCase__ ) A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 32, 32, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = """cpu""" # ensure determinism for the device-dependent torch.Generator A_ = self.get_dummy_components(class_cond=UpperCamelCase__ ) A_ = ConsistencyModelPipeline(**UpperCamelCase__ ) A_ = pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_dummy_inputs(UpperCamelCase__ ) A_ = 0 A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 32, 32, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def snake_case_ ( self ) -> Tuple: '''simple docstring''' A_ = """cpu""" # ensure determinism for the device-dependent torch.Generator A_ = self.get_dummy_components() A_ = ConsistencyModelPipeline(**UpperCamelCase__ ) A_ = pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_dummy_inputs(UpperCamelCase__ ) A_ = 1 A_ = None A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 32, 32, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' A_ = """cpu""" # ensure determinism for the device-dependent torch.Generator A_ = self.get_dummy_components(class_cond=UpperCamelCase__ ) A_ = ConsistencyModelPipeline(**UpperCamelCase__ ) A_ = pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_dummy_inputs(UpperCamelCase__ ) A_ = 1 A_ = None A_ = 0 A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 32, 32, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case_ ( self , UpperCamelCase__=0 , UpperCamelCase__=False , UpperCamelCase__="cpu" , UpperCamelCase__=torch.floataa , UpperCamelCase__=(1, 3, 64, 64) ) -> Any: '''simple docstring''' A_ = torch.manual_seed(UpperCamelCase__ ) A_ = { """num_inference_steps""": None, """timesteps""": [22, 0], """class_labels""": 0, """generator""": generator, """output_type""": """np""", } if get_fixed_latents: A_ = self.get_fixed_latents(seed=UpperCamelCase__ , device=UpperCamelCase__ , dtype=UpperCamelCase__ , shape=UpperCamelCase__ ) A_ = latents return inputs def snake_case_ ( self , UpperCamelCase__=0 , UpperCamelCase__="cpu" , UpperCamelCase__=torch.floataa , UpperCamelCase__=(1, 3, 64, 64) ) -> Dict: '''simple docstring''' if type(UpperCamelCase__ ) == str: A_ = torch.device(UpperCamelCase__ ) A_ = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A_ = randn_tensor(UpperCamelCase__ , generator=UpperCamelCase__ , device=UpperCamelCase__ , dtype=UpperCamelCase__ ) return latents def snake_case_ ( self ) -> List[Any]: '''simple docstring''' A_ = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) A_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) A_ = ConsistencyModelPipeline(unet=UpperCamelCase__ , scheduler=UpperCamelCase__ ) pipe.to(torch_device=UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_inputs() A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 64, 64, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.0888, 0.0881, 0.0666, 0.0479, 0.0292, 0.0195, 0.0201, 0.0163, 0.0254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' A_ = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) A_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) A_ = ConsistencyModelPipeline(unet=UpperCamelCase__ , scheduler=UpperCamelCase__ ) pipe.to(torch_device=UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_inputs() A_ = 1 A_ = None A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 64, 64, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.0340, 0.0152, 0.0063, 0.0267, 0.0221, 0.0107, 0.0416, 0.0186, 0.0217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) A_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) A_ = ConsistencyModelPipeline(unet=UpperCamelCase__ , scheduler=UpperCamelCase__ ) pipe.to(torch_device=UpperCamelCase__ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_inputs(get_fixed_latents=UpperCamelCase__ , device=UpperCamelCase__ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=UpperCamelCase__ , enable_math=UpperCamelCase__ , enable_mem_efficient=UpperCamelCase__ ): A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 64, 64, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.1875, 0.1428, 0.1289, 0.2151, 0.2092, 0.1477, 0.1877, 0.1641, 0.1353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def snake_case_ ( self ) -> Optional[int]: '''simple docstring''' A_ = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) A_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) A_ = ConsistencyModelPipeline(unet=UpperCamelCase__ , scheduler=UpperCamelCase__ ) pipe.to(torch_device=UpperCamelCase__ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) A_ = self.get_inputs(get_fixed_latents=UpperCamelCase__ , device=UpperCamelCase__ ) A_ = 1 A_ = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=UpperCamelCase__ , enable_math=UpperCamelCase__ , enable_mem_efficient=UpperCamelCase__ ): A_ = pipe(**UpperCamelCase__ ).images assert image.shape == (1, 64, 64, 3) A_ = image[0, -3:, -3:, -1] A_ = np.array([0.1663, 0.1948, 0.2275, 0.1680, 0.1204, 0.1245, 0.1858, 0.1338, 0.2095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
162
'''simple docstring''' import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params __lowerCamelCase = getLogger(__name__) __lowerCamelCase = '''cuda''' if torch.cuda.is_available() else '''cpu''' def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ = 8, UpperCAmelCase__ = DEFAULT_DEVICE, UpperCAmelCase__=False, UpperCAmelCase__="summarization", UpperCAmelCase__=None, **UpperCAmelCase__, ) -> Dict: A_ = Path(UpperCAmelCase__ ).open("""w""", encoding="""utf-8""" ) A_ = str(UpperCAmelCase__ ) A_ = AutoModelForSeqaSeqLM.from_pretrained(UpperCAmelCase__ ).to(UpperCAmelCase__ ) if fpaa: A_ = model.half() A_ = AutoTokenizer.from_pretrained(UpperCAmelCase__ ) logger.info(F'''Inferred tokenizer type: {tokenizer.__class__}''' ) # if this is wrong, check config.model_type. A_ = time.time() # update config with task specific params use_task_specific_params(UpperCAmelCase__, UpperCAmelCase__ ) if prefix is None: A_ = prefix or getattr(model.config, """prefix""", """""" ) or """""" for examples_chunk in tqdm(list(chunks(UpperCAmelCase__, UpperCAmelCase__ ) ) ): A_ = [prefix + text for text in examples_chunk] A_ = tokenizer(UpperCAmelCase__, return_tensors="""pt""", truncation=UpperCAmelCase__, padding="""longest""" ).to(UpperCAmelCase__ ) A_ = model.generate( input_ids=batch.input_ids, attention_mask=batch.attention_mask, **UpperCAmelCase__, ) A_ = tokenizer.batch_decode(UpperCAmelCase__, skip_special_tokens=UpperCAmelCase__, clean_up_tokenization_spaces=UpperCAmelCase__ ) for hypothesis in dec: fout.write(hypothesis + """\n""" ) fout.flush() fout.close() A_ = int(time.time() - start_time ) # seconds A_ = len(UpperCAmelCase__ ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs, 4 )} def UpperCAmelCase__ ( ) -> Optional[int]: return datetime.datetime.now().strftime("""%Y-%m-%d %H:%M:%S""" ) def UpperCAmelCase__ ( UpperCAmelCase__=True ) -> Any: A_ = argparse.ArgumentParser() parser.add_argument("""model_name""", type=UpperCAmelCase__, help="""like facebook/bart-large-cnn,t5-base, etc.""" ) parser.add_argument("""input_path""", type=UpperCAmelCase__, help="""like cnn_dm/test.source""" ) parser.add_argument("""save_path""", type=UpperCAmelCase__, help="""where to save summaries""" ) parser.add_argument("""--reference_path""", type=UpperCAmelCase__, required=UpperCAmelCase__, help="""like cnn_dm/test.target""" ) parser.add_argument("""--score_path""", type=UpperCAmelCase__, required=UpperCAmelCase__, default="""metrics.json""", help="""where to save metrics""" ) parser.add_argument("""--device""", type=UpperCAmelCase__, required=UpperCAmelCase__, default=UpperCAmelCase__, help="""cuda, cuda:1, cpu etc.""" ) parser.add_argument( """--prefix""", type=UpperCAmelCase__, required=UpperCAmelCase__, default=UpperCAmelCase__, help="""will be added to the begininng of src examples""" ) parser.add_argument("""--task""", type=UpperCAmelCase__, default="""summarization""", help="""used for task_specific_params + metrics""" ) parser.add_argument("""--bs""", type=UpperCAmelCase__, default=8, required=UpperCAmelCase__, help="""batch size""" ) parser.add_argument( """--n_obs""", type=UpperCAmelCase__, default=-1, required=UpperCAmelCase__, help="""How many observations. Defaults to all.""" ) parser.add_argument("""--fp16""", action="""store_true""" ) parser.add_argument("""--dump-args""", action="""store_true""", help="""print the custom hparams with the results""" ) parser.add_argument( """--info""", nargs="""?""", type=UpperCAmelCase__, const=datetime_now(), help=( """use in conjunction w/ --dump-args to print with the results whatever other info you'd like, e.g.""" """ lang=en-ru. If no value is passed, the current datetime string will be used.""" ), ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate A_ , A_ = parser.parse_known_args() A_ = parse_numeric_n_bool_cl_kwargs(UpperCAmelCase__ ) if parsed_args and verbose: print(F'''parsed the following generate kwargs: {parsed_args}''' ) A_ = [""" """ + x.rstrip() if """t5""" in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: A_ = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=UpperCAmelCase__ ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F'''score_path {args.score_path} will be overwritten unless you type ctrl-c.''' ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError("""Can't mix --fp16 and --device cpu""" ) A_ = generate_summaries_or_translations( UpperCAmelCase__, args.save_path, args.model_name, batch_size=args.bs, device=args.device, fpaa=args.fpaa, task=args.task, prefix=args.prefix, **UpperCAmelCase__, ) if args.reference_path is None: return {} # Compute scores A_ = calculate_bleu if """translation""" in args.task else calculate_rouge A_ = [x.rstrip() for x in open(args.save_path ).readlines()] A_ = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(UpperCAmelCase__ )] A_ = score_fn(UpperCAmelCase__, UpperCAmelCase__ ) scores.update(UpperCAmelCase__ ) if args.dump_args: scores.update(UpperCAmelCase__ ) if args.info: A_ = args.info if verbose: print(UpperCAmelCase__ ) if args.score_path is not None: json.dump(UpperCAmelCase__, open(args.score_path, """w""" ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
162
1
import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[Any]=100 , SCREAMING_SNAKE_CASE__ : Optional[Any]=13 , SCREAMING_SNAKE_CASE__ : Dict=30 , SCREAMING_SNAKE_CASE__ : Any=2 , SCREAMING_SNAKE_CASE__ : str=3 , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : Dict=32 , SCREAMING_SNAKE_CASE__ : Optional[Any]=5 , SCREAMING_SNAKE_CASE__ : Optional[Any]=4 , SCREAMING_SNAKE_CASE__ : Optional[Any]=37 , SCREAMING_SNAKE_CASE__ : Any="gelu" , SCREAMING_SNAKE_CASE__ : List[Any]=0.1 , SCREAMING_SNAKE_CASE__ : int=0.1 , SCREAMING_SNAKE_CASE__ : Tuple=10 , SCREAMING_SNAKE_CASE__ : List[str]=0.02 , SCREAMING_SNAKE_CASE__ : Any=3 , ) -> Optional[int]: lowerCAmelCase__ = parent lowerCAmelCase__ = vocab_size lowerCAmelCase__ = batch_size lowerCAmelCase__ = image_size lowerCAmelCase__ = patch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = is_training lowerCAmelCase__ = use_labels lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_act lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = type_sequence_label_size lowerCAmelCase__ = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ = (image_size // patch_size) ** 2 lowerCAmelCase__ = num_patches + 1 def a ( self : List[Any] ) -> int: lowerCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ = BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE__ , initializer_range=self.initializer_range , ) return config, pixel_values, labels def a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : int ) -> int: lowerCAmelCase__ = FlaxBeitModel(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple ) -> List[Any]: lowerCAmelCase__ = FlaxBeitForMaskedImageModeling(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def a ( self : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> int: lowerCAmelCase__ = self.type_sequence_label_size lowerCAmelCase__ = FlaxBeitForImageClassification(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ = 1 lowerCAmelCase__ = FlaxBeitForImageClassification(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) def a ( self : List[Any] ) -> Optional[int]: lowerCAmelCase__ = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) = config_and_inputs lowerCAmelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class __lowerCamelCase ( UpperCamelCase__ , unittest.TestCase ): """simple docstring""" snake_case__ = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def a ( self : Optional[int] ) -> None: lowerCAmelCase__ = FlaxBeitModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ , hidden_size=37 ) def a ( self : str ) -> Optional[Any]: self.config_tester.run_common_tests() def a ( self : str ) -> Any: lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ = [*signature.parameters.keys()] lowerCAmelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] ) -> int: lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowerCAmelCase__ = self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) @jax.jit def model_jitted(SCREAMING_SNAKE_CASE__ : Tuple , **SCREAMING_SNAKE_CASE__ : Optional[int] ): return model(pixel_values=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) with self.subTest("JIT Enabled" ): lowerCAmelCase__ = model_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): lowerCAmelCase__ = model_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , len(SCREAMING_SNAKE_CASE__ ) ) for jitted_output, output in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): self.assertEqual(jitted_output.shape , output.shape ) def a ( self : Optional[int] ) -> str: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[Any] ) -> Any: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*SCREAMING_SNAKE_CASE__ ) def a ( self : str ) -> List[Any]: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE__ ) @slow def a ( self : List[str] ) -> Any: for model_class_name in self.all_model_classes: lowerCAmelCase__ = model_class_name.from_pretrained("microsoft/beit-base-patch16-224" ) lowerCAmelCase__ = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def _A ( ): """simple docstring""" lowerCAmelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @require_flax class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self : Any ) -> Union[str, Any]: return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def a ( self : List[Any] ) -> Optional[Any]: lowerCAmelCase__ = FlaxBeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="np" ).pixel_values # prepare bool_masked_pos lowerCAmelCase__ = np.ones((1, 196) , dtype=SCREAMING_SNAKE_CASE__ ) # forward pass lowerCAmelCase__ = model(pixel_values=SCREAMING_SNAKE_CASE__ , bool_masked_pos=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = outputs.logits # verify the logits lowerCAmelCase__ = (1, 196, 8_192) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = np.array( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-2 ) ) @slow def a ( self : int ) -> List[Any]: lowerCAmelCase__ = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="np" ) # forward pass lowerCAmelCase__ = model(**SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = outputs.logits # verify the logits lowerCAmelCase__ = (1, 1_000) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = np.array([-1.2_385, -1.0_987, -1.0_108] ) self.assertTrue(np.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) lowerCAmelCase__ = 281 self.assertEqual(logits.argmax(-1 ).item() , SCREAMING_SNAKE_CASE__ ) @slow def a ( self : Optional[int] ) -> Optional[int]: lowerCAmelCase__ = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="np" ) # forward pass lowerCAmelCase__ = model(**SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = outputs.logits # verify the logits lowerCAmelCase__ = (1, 21_841) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = np.array([1.6_881, -0.2_787, 0.5_901] ) self.assertTrue(np.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) lowerCAmelCase__ = 2_396 self.assertEqual(logits.argmax(-1 ).item() , SCREAMING_SNAKE_CASE__ )
221
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" snake_case__ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def a ( self : str , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Any: lowerCAmelCase__ = hf_hub_download( repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) lowerCAmelCase__ = VideoClassificationPipeline(model=SCREAMING_SNAKE_CASE__ , image_processor=SCREAMING_SNAKE_CASE__ , top_k=2 ) lowerCAmelCase__ = [ example_video_filepath, "https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4", ] return video_classifier, examples def a ( self : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Tuple: for example in examples: lowerCAmelCase__ = video_classifier(SCREAMING_SNAKE_CASE__ ) self.assertEqual( SCREAMING_SNAKE_CASE__ , [ {"score": ANY(SCREAMING_SNAKE_CASE__ ), "label": ANY(SCREAMING_SNAKE_CASE__ )}, {"score": ANY(SCREAMING_SNAKE_CASE__ ), "label": ANY(SCREAMING_SNAKE_CASE__ )}, ] , ) @require_torch def a ( self : Dict ) -> Optional[Any]: lowerCAmelCase__ = "hf-internal-testing/tiny-random-VideoMAEForVideoClassification" lowerCAmelCase__ = VideoMAEFeatureExtractor( size={"shortest_edge": 10} , crop_size={"height": 10, "width": 10} ) lowerCAmelCase__ = pipeline( "video-classification" , model=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ , frame_sampling_rate=4 ) lowerCAmelCase__ = hf_hub_download(repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) lowerCAmelCase__ = video_classifier(SCREAMING_SNAKE_CASE__ , top_k=2 ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ , decimals=4 ) , [{"score": 0.5_199, "label": "LABEL_0"}, {"score": 0.4_801, "label": "LABEL_1"}] , ) lowerCAmelCase__ = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ , decimals=4 ) , [ [{"score": 0.5_199, "label": "LABEL_0"}, {"score": 0.4_801, "label": "LABEL_1"}], [{"score": 0.5_199, "label": "LABEL_0"}, {"score": 0.4_801, "label": "LABEL_1"}], ] , ) @require_tf def a ( self : Optional[Any] ) -> Optional[int]: pass
221
1
"""simple docstring""" import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def UpperCamelCase ( ) ->int: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(UpperCAmelCase ): requests.request("GET" , "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET" , "https://huggingface.co" , timeout=1.0 ) @pytest.mark.integration def UpperCamelCase ( ) ->Tuple: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET" , "https://huggingface.co" ) def UpperCamelCase ( ) ->Optional[Any]: """simple docstring""" with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(UpperCAmelCase ): http_head("https://huggingface.co" )
243
import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __lowerCAmelCase : List[Any] = get_tests_dir('fixtures/test_sentencepiece_with_bytefallback.model') @require_sentencepiece @require_tokenizers class snake_case__ (_UpperCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = GPTSwaTokenizer SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = True SCREAMING_SNAKE_CASE_ : Any = False def __UpperCAmelCase ( self : Tuple ) -> Any: super().setUp() # We have a SentencePiece fixture for testing a = GPTSwaTokenizer(__lowerCamelCase , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCAmelCase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Any: a = "This is a test" a = "This is a test" return input_text, output_text def __UpperCAmelCase ( self : List[Any] ) -> List[str]: a = "<s>" a = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def __UpperCAmelCase ( self : List[Any] ) -> int: a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(__lowerCamelCase ) , 20_00 ) def __UpperCAmelCase ( self : Optional[int] ) -> Dict: self.assertEqual(self.get_tokenizer().vocab_size , 20_00 ) def __UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: a = GPTSwaTokenizer(__lowerCamelCase ) a = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [4_65, 2_87, 2_65, 6_31, 8_42] ) a = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( __lowerCamelCase , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on a = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [2_62, 2_72, 15_25, 2_86, 2_71, 2_68, 60, 9_16, 6_33, 6_33, 6_33, 2_59, 2_66, 3_01, 2_87, 3_84, 3_67, 2_63, 1_98, 1_72, 2_60] , ) a = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) # fmt: off self.assertListEqual( __lowerCamelCase , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def __UpperCAmelCase ( self : List[Any] ) -> str: a = GPTSwaTokenizer(__lowerCamelCase ) a = ["This is a test", "I was born in 92000, and this is falsé."] a = [ [4_65, 2_87, 2_65, 6_31, 8_42], [2_62, 2_72, 15_25, 2_86, 2_71, 2_68, 60, 9_16, 6_33, 6_33, 6_33, 2_59, 2_66, 3_01, 2_87, 3_84, 3_67, 2_63, 1_98, 1_72, 2_60], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(__lowerCamelCase , __lowerCamelCase ): self.assertListEqual(tokenizer.encode_fast(__lowerCamelCase ) , __lowerCamelCase ) # Test that decode_fast returns the input text for text, token_ids in zip(__lowerCamelCase , __lowerCamelCase ): self.assertEqual(tokenizer.decode_fast(__lowerCamelCase ) , __lowerCamelCase ) @slow def __UpperCAmelCase ( self : Optional[int] ) -> List[Any]: a = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off a = {"input_ids": [[6_34_23, 5, 68_11, 1_49_54, 2_82, 8_16, 38_21, 6_34_66, 6_34_25, 6_34_62, 18, 6_39_78, 6_78, 3_01, 13_20, 6_34_23, 6_34_55, 6_34_58, 18, 6_39_82, 42_46, 39_40, 19_01, 4_77_89, 55_47, 1_89_94], [1_96_30, 11_00, 6_34_46, 13_42, 6_33, 5_44, 44_88, 5_93, 51_02, 24_16, 6_34_95, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [16_52, 4_28, 2_68, 19_36, 5_15, 2_68, 5_85_93, 2_24_13, 91_06, 5_46, 2_68, 3_32_13, 6_39_79, 6_98, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_51_30, 6_34_50, 9_24, 6_34_49, 22_49, 40_62, 15_58, 3_18, 6_35_04, 2_14_98, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_09, 3_77, 28_27, 25_59, 3_32, 65_75, 6_34_43, 2_68_01, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="AI-Sweden/gpt-sw3-126m" , sequences=__lowerCamelCase , )
107
0
import logging from transformers.configuration_utils import PretrainedConfig UpperCamelCase__ = logging.getLogger(__name__) class lowerCamelCase_ ( __snake_case ): lowerCAmelCase__ = 'masked_bert' def __init__( self : Any , _A : List[str]=30_522 , _A : Dict=768 , _A : Optional[Any]=12 , _A : int=12 , _A : str=3_072 , _A : int="gelu" , _A : Optional[Any]=0.1 , _A : str=0.1 , _A : Optional[Any]=512 , _A : Tuple=2 , _A : Optional[Any]=0.0_2 , _A : Dict=1e-12 , _A : Optional[Any]=0 , _A : Tuple="topK" , _A : Optional[int]="constant" , _A : Optional[int]=0.0 , **_A : List[Any] , ): '''simple docstring''' super().__init__(pad_token_id=_A , **_A ) UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : int = hidden_act UpperCAmelCase__ : List[str] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Dict = type_vocab_size UpperCAmelCase__ : Dict = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : List[str] = pruning_method UpperCAmelCase__ : Optional[int] = mask_init UpperCAmelCase__ : Any = mask_scale
362
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : int , *_A : Tuple , **_A : Union[str, Any] ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Union[str, Any] , *_A : List[Any] , **_A : Any ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : int , *_A : Union[str, Any] , **_A : Optional[Any] ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : Any , *_A : List[str] , **_A : Tuple ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Tuple , *_A : Tuple , **_A : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : List[Any] , *_A : List[str] , **_A : Dict ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : Dict , *_A : Any , **_A : int ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : List[Any] , *_A : List[Any] , **_A : Optional[int] ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : int , *_A : Dict , **_A : Optional[Any] ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : List[Any] , *_A : Optional[int] , **_A : Optional[Any] ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Dict , *_A : Any , **_A : Tuple ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : int , *_A : Union[str, Any] , **_A : Dict ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : List[Any] , *_A : Optional[int] , **_A : Dict ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Dict , *_A : str , **_A : Dict ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Union[str, Any] , *_A : Optional[int] , **_A : int ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) class lowerCamelCase_ ( metaclass=__a ): lowerCAmelCase__ = ['torch', 'transformers', 'onnx'] def __init__( self : Optional[Any] , *_A : Union[str, Any] , **_A : Dict ): '''simple docstring''' requires_backends(self , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : List[str] , *_A : str , **_A : List[str] ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] ) @classmethod def lowercase_ ( cls : Dict , *_A : str , **_A : Any ): '''simple docstring''' requires_backends(cls , ['''torch''', '''transformers''', '''onnx'''] )
299
0
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("0.8.3"): raise Exception("requires gluonnlp == 0.8.3") if version.parse(mx.__version__) != version.parse("1.5.0"): raise Exception("requires mxnet == 1.5.0") logging.set_verbosity_info() _lowercase : Union[str, Any] =logging.get_logger(__name__) _lowercase : Optional[Any] ="The Nymphenburg Palace is a beautiful palace in Munich!" def lowerCAmelCase_ ( _lowercase : str , _lowercase : str) -> Tuple: """simple docstring""" a__ : List[Any] = { """attention_cell""": """multi_head""", """num_layers""": 4, """units""": 1024, """hidden_size""": 768, """max_length""": 512, """num_heads""": 8, """scaled""": True, """dropout""": 0.1, """use_residual""": True, """embed_size""": 1024, """embed_dropout""": 0.1, """word_embed""": None, """layer_norm_eps""": 1e-5, """token_type_vocab_size""": 2, } a__ : List[str] = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py a__ : Union[str, Any] = BERTEncoder( attention_cell=predefined_args["""attention_cell"""] , num_layers=predefined_args["""num_layers"""] , units=predefined_args["""units"""] , hidden_size=predefined_args["""hidden_size"""] , max_length=predefined_args["""max_length"""] , num_heads=predefined_args["""num_heads"""] , scaled=predefined_args["""scaled"""] , dropout=predefined_args["""dropout"""] , output_attention=_lowercase , output_all_encodings=_lowercase , use_residual=predefined_args["""use_residual"""] , activation=predefined_args.get("""activation""" , """gelu""") , layer_norm_eps=predefined_args.get("""layer_norm_eps""" , _lowercase) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later a__ : Dict = """openwebtext_ccnews_stories_books_cased""" # Specify download folder to Gluonnlp's vocab a__ : int = os.path.join(get_home_dir() , """models""") a__ : Tuple = _load_vocab(_lowercase , _lowercase , _lowercase , cls=_lowercase) a__ : List[Any] = nlp.model.BERTModel( _lowercase , len(_lowercase) , units=predefined_args["""units"""] , embed_size=predefined_args["""embed_size"""] , embed_dropout=predefined_args["""embed_dropout"""] , word_embed=predefined_args["""word_embed"""] , use_pooler=_lowercase , use_token_type_embed=_lowercase , token_type_vocab_size=predefined_args["""token_type_vocab_size"""] , use_classifier=_lowercase , use_decoder=_lowercase , ) original_bort.load_parameters(_lowercase , cast_dtype=_lowercase , ignore_extra=_lowercase) a__ : Optional[Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 a__ : Optional[int] = { """architectures""": ["""BertForMaskedLM"""], """attention_probs_dropout_prob""": predefined_args["""dropout"""], """hidden_act""": """gelu""", """hidden_dropout_prob""": predefined_args["""dropout"""], """hidden_size""": predefined_args["""embed_size"""], """initializer_range""": 0.02, """intermediate_size""": predefined_args["""hidden_size"""], """layer_norm_eps""": predefined_args["""layer_norm_eps"""], """max_position_embeddings""": predefined_args["""max_length"""], """model_type""": """bort""", """num_attention_heads""": predefined_args["""num_heads"""], """num_hidden_layers""": predefined_args["""num_layers"""], """pad_token_id""": 1, # 2 = BERT, 1 = RoBERTa """type_vocab_size""": 1, # 2 = BERT, 1 = RoBERTa """vocab_size""": len(_lowercase), } a__ : Tuple = BertConfig.from_dict(_lowercase) a__ : Any = BertForMaskedLM(_lowercase) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(_lowercase : Any) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy())) # Check param shapes and map new HF param back def check_and_map_params(_lowercase : Dict , _lowercase : Union[str, Any]): a__ : Tuple = hf_param.shape a__ : List[str] = to_torch(params[gluon_param]) a__ : Dict = gluon_param.shape assert ( shape_hf == shape_gluon ), F'''The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers''' return gluon_param a__ : Optional[Any] = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , """word_embed.0.weight""") a__ : Optional[int] = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , """encoder.position_weight""") a__ : Optional[Any] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , """encoder.layer_norm.beta""") a__ : Union[str, Any] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , """encoder.layer_norm.gamma""") # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) a__ : Dict = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data) for i in range(hf_bort_config.num_hidden_layers): a__ : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention a__ : BertSelfAttention = layer.attention.self a__ : str = check_and_map_params( self_attn.key.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_key.bias''') a__ : List[Any] = check_and_map_params( self_attn.key.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_key.weight''') a__ : str = check_and_map_params( self_attn.query.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_query.bias''') a__ : Any = check_and_map_params( self_attn.query.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_query.weight''') a__ : List[Any] = check_and_map_params( self_attn.value.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_value.bias''') a__ : Any = check_and_map_params( self_attn.value.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_value.weight''') # self attention output a__ : BertSelfOutput = layer.attention.output a__ : str = check_and_map_params( self_output.dense.bias , F'''encoder.transformer_cells.{i}.proj.bias''') a__ : List[Any] = check_and_map_params( self_output.dense.weight , F'''encoder.transformer_cells.{i}.proj.weight''') a__ : Any = check_and_map_params( self_output.LayerNorm.bias , F'''encoder.transformer_cells.{i}.layer_norm.beta''') a__ : Any = check_and_map_params( self_output.LayerNorm.weight , F'''encoder.transformer_cells.{i}.layer_norm.gamma''') # intermediate a__ : BertIntermediate = layer.intermediate a__ : int = check_and_map_params( intermediate.dense.bias , F'''encoder.transformer_cells.{i}.ffn.ffn_1.bias''') a__ : int = check_and_map_params( intermediate.dense.weight , F'''encoder.transformer_cells.{i}.ffn.ffn_1.weight''') # output a__ : BertOutput = layer.output a__ : List[Any] = check_and_map_params( bert_output.dense.bias , F'''encoder.transformer_cells.{i}.ffn.ffn_2.bias''') a__ : str = check_and_map_params( bert_output.dense.weight , F'''encoder.transformer_cells.{i}.ffn.ffn_2.weight''') a__ : int = check_and_map_params( bert_output.LayerNorm.bias , F'''encoder.transformer_cells.{i}.ffn.layer_norm.beta''') a__ : Optional[Any] = check_and_map_params( bert_output.LayerNorm.weight , F'''encoder.transformer_cells.{i}.ffn.layer_norm.gamma''') # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models a__ : Dict = RobertaTokenizer.from_pretrained("""roberta-base""") a__ : int = tokenizer.encode_plus(_lowercase)["""input_ids"""] # Get gluon output a__ : Optional[int] = mx.nd.array([input_ids]) a__ : Dict = original_bort(inputs=_lowercase , token_types=[]) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(_lowercase) a__ : Dict = BertModel.from_pretrained(_lowercase) hf_bort_model.eval() a__ : Any = tokenizer.encode_plus(_lowercase , return_tensors="""pt""") a__ : Dict = hf_bort_model(**_lowercase)[0] a__ : str = output_gluon[0].asnumpy() a__ : str = output_hf[0].detach().numpy() a__ : Tuple = np.max(np.abs(hf_layer - gluon_layer)).item() a__ : int = np.allclose(_lowercase , _lowercase , atol=1e-3) if success: print("""✔️ Both model do output the same tensors""") else: print("""❌ Both model do **NOT** output the same tensors""") print("""Absolute difference is:""" , _lowercase) if __name__ == "__main__": _lowercase : str =argparse.ArgumentParser() # Required parameters parser.add_argument( "--bort_checkpoint_path", default=None, type=str, required=True, help="Path the official Bort params file." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _lowercase : Optional[int] =parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
170
import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def lowerCAmelCase_ ( _lowercase : Dict , _lowercase : str , _lowercase : str , _lowercase : Optional[Any]=1024) -> List[Any]: """simple docstring""" a__ , a__ : Optional[int] = [], [] a__ : Union[str, Any] = list(zip(_lowercase , _lowercase)) a__ , a__ : List[Any] = sorted_examples[0] def is_too_big(_lowercase : Tuple): return tok(_lowercase , return_tensors="""pt""").input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:]): a__ : Tuple = new_src + """ """ + src a__ : Any = new_tgt + """ """ + tgt if is_too_big(_lowercase) or is_too_big(_lowercase): # cant fit, finalize example finished_src.append(_lowercase) finished_tgt.append(_lowercase) a__ , a__ : List[Any] = src, tgt else: # can fit, keep adding a__ , a__ : Tuple = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(_lowercase) finished_tgt.append(_lowercase) return finished_src, finished_tgt def lowerCAmelCase_ ( _lowercase : str , _lowercase : Path , _lowercase : Any , _lowercase : str) -> Tuple: """simple docstring""" a__ : Any = Path(_lowercase) save_path.mkdir(exist_ok=_lowercase) for split in ["train"]: a__ , a__ : List[Any] = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' a__ : Dict = [x.rstrip() for x in Path(_lowercase).open().readlines()] a__ : Optional[Any] = [x.rstrip() for x in Path(_lowercase).open().readlines()] a__ , a__ : List[Any] = pack_examples(_lowercase , _lowercase , _lowercase , _lowercase) print(F'''packed {split} split from {len(_lowercase)} examples -> {len(_lowercase)}.''') Path(save_path / F'''{split}.source''').open("""w""").write("""\n""".join(_lowercase)) Path(save_path / F'''{split}.target''').open("""w""").write("""\n""".join(_lowercase)) for split in ["val", "test"]: a__ , a__ : Any = data_dir / F'''{split}.source''', data_dir / F'''{split}.target''' shutil.copyfile(_lowercase , save_path / F'''{split}.source''') shutil.copyfile(_lowercase , save_path / F'''{split}.target''') def lowerCAmelCase_ ( ) -> Optional[int]: """simple docstring""" a__ : Tuple = argparse.ArgumentParser() parser.add_argument("""--tok_name""" , type=_lowercase , help="""like facebook/bart-large-cnn,t5-base, etc.""") parser.add_argument("""--max_seq_len""" , type=_lowercase , default=128) parser.add_argument("""--data_dir""" , type=_lowercase) parser.add_argument("""--save_path""" , type=_lowercase) a__ : List[Any] = parser.parse_args() a__ : List[Any] = AutoTokenizer.from_pretrained(args.tok_name) return pack_data_dir(_lowercase , Path(args.data_dir) , args.max_seq_len , args.save_path) if __name__ == "__main__": packer_cli()
170
1
import math def snake_case__ ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE_ , 2 ) - a def snake_case__ ( SCREAMING_SNAKE_CASE_ : float ): '''simple docstring''' return 2 * x def snake_case__ ( SCREAMING_SNAKE_CASE_ : float ): '''simple docstring''' lowercase__ : Union[str, Any] = 2.0 while start <= a: lowercase__ : Tuple = math.pow(SCREAMING_SNAKE_CASE_ , 2 ) return start def snake_case__ ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : int = 9_999 , SCREAMING_SNAKE_CASE_ : float = 0.00_0000_0000_0001 ): '''simple docstring''' if a < 0: raise ValueError('math domain error' ) lowercase__ : Optional[int] = get_initial_point(SCREAMING_SNAKE_CASE_ ) for _ in range(SCREAMING_SNAKE_CASE_ ): lowercase__ : Dict = value lowercase__ : List[Any] = value - fx(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) / fx_derivative(SCREAMING_SNAKE_CASE_ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
216
import math import sys def snake_case__ ( SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' if number != int(SCREAMING_SNAKE_CASE_ ): raise ValueError('the value of input must be a natural number' ) if number < 0: raise ValueError('the value of input must not be a negative number' ) if number == 0: return 1 lowercase__ : Tuple = [-1] * (number + 1) lowercase__ : Tuple = 0 for i in range(1 , number + 1 ): lowercase__ : Tuple = sys.maxsize lowercase__ : str = int(math.sqrt(SCREAMING_SNAKE_CASE_ ) ) for j in range(1 , root + 1 ): lowercase__ : List[Any] = 1 + answers[i - (j**2)] lowercase__ : str = min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ : List[str] = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
216
1
'''simple docstring''' from importlib import import_module from .logging import get_logger lowerCAmelCase__ = get_logger(__name__) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : Dict=None ): __lowercase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__''' ): setattr(self ,lowercase__ ,getattr(lowercase__ ,lowercase__ ) ) __lowercase = module._original_module if isinstance(lowercase__ ,_PatchedModuleObj ) else module class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : str = [] def __init__( self : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : str ,lowercase__ : Dict ,lowercase__ : Optional[int]=None ): __lowercase = obj __lowercase = target __lowercase = new __lowercase = target.split('''.''' )[0] __lowercase = {} __lowercase = attrs or [] def __enter__( self : Tuple ): *__lowercase , __lowercase = self.target.split('''.''' ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowercase__ ) ): try: __lowercase = import_module('''.'''.join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): __lowercase = getattr(self.obj ,lowercase__ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowercase__ ,_PatchedModuleObj ) and obj_attr._original_module is submodule) ): __lowercase = obj_attr # patch at top level setattr(self.obj ,lowercase__ ,_PatchedModuleObj(lowercase__ ,attrs=self.attrs ) ) __lowercase = getattr(self.obj ,lowercase__ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowercase__ ,lowercase__ ,_PatchedModuleObj(getattr(lowercase__ ,lowercase__ ,lowercase__ ) ,attrs=self.attrs ) ) __lowercase = getattr(lowercase__ ,lowercase__ ) # finally set the target attribute setattr(lowercase__ ,lowercase__ ,self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: __lowercase = getattr(import_module('''.'''.join(lowercase__ ) ) ,lowercase__ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj ,lowercase__ ) is attr_value: __lowercase = getattr(self.obj ,lowercase__ ) setattr(self.obj ,lowercase__ ,self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" __lowercase = globals()['''__builtins__'''][target_attr] setattr(self.obj ,lowercase__ ,self.new ) else: raise RuntimeError(F"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self : Optional[Any] ,*lowercase__ : int ): for attr in list(self.original ): setattr(self.obj ,lowercase__ ,self.original.pop(lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): self.__enter__() self._active_patches.append(self ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
104
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version UpperCAmelCase_ = get_logger(__name__) class lowerCamelCase__: UpperCAmelCase__ : List[Any] = 'dummy_data' UpperCAmelCase__ : str = 'datasets' UpperCAmelCase__ : Tuple = False def __init__( self: Optional[Any] , UpperCamelCase_: str , UpperCamelCase_: str , UpperCamelCase_: Union[Version, str] , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: bool = False , UpperCamelCase_: bool = True , UpperCamelCase_: Optional[List[Callable]] = None , ): __lowerCamelCase = 0 __lowerCamelCase = dataset_name __lowerCamelCase = cache_dir __lowerCamelCase = use_local_dummy_data __lowerCamelCase = config # download_callbacks take a single url as input __lowerCamelCase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root __lowerCamelCase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general __lowerCamelCase = str(UpperCamelCase_ ) # to be downloaded __lowerCamelCase = None __lowerCamelCase = None @property def lowerCAmelCase__ ( self: List[Any] ): if self._dummy_file is None: __lowerCamelCase = self.download_dummy_data() return self._dummy_file @property def lowerCAmelCase__ ( self: str ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join("""dummy""" , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join("""dummy""" , self.version_name ) @property def lowerCAmelCase__ ( self: Optional[Any] ): return os.path.join(self.dummy_data_folder , """dummy_data.zip""" ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) __lowerCamelCase = cached_path( UpperCamelCase_ , cache_dir=self.cache_dir , extract_compressed_file=UpperCamelCase_ , force_extract=UpperCamelCase_ ) return os.path.join(UpperCamelCase_ , self.dummy_file_name ) @property def lowerCAmelCase__ ( self: Optional[Any] ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def lowerCAmelCase__ ( self: Tuple ): if self._bucket_url is None: __lowerCamelCase = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , """/""" ) ) return self._bucket_url @property def lowerCAmelCase__ ( self: str ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , """/""" ).split("""/""" )[:-1] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Dict , *UpperCamelCase_: str ): if self.load_existing_dummy_data: # dummy data is downloaded and tested __lowerCamelCase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned __lowerCamelCase = self.dummy_file_name # special case when data_url is a dict if isinstance(UpperCamelCase_ , UpperCamelCase_ ): return self.create_dummy_data_dict(UpperCamelCase_ , UpperCamelCase_ ) elif isinstance(UpperCamelCase_ , (list, tuple) ): return self.create_dummy_data_list(UpperCamelCase_ , UpperCamelCase_ ) else: return self.create_dummy_data_single(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Optional[Any] , *UpperCamelCase_: str ): return self.download_and_extract(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: str ): return self.download_and_extract(UpperCamelCase_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: int , *UpperCamelCase_: List[str] , **UpperCamelCase_: str ): return path def lowerCAmelCase__ ( self: Dict ): return {} def lowerCAmelCase__ ( self: str , UpperCamelCase_: List[Any] , UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(UpperCamelCase_ , UpperCamelCase_ ): for single_url in single_urls: download_callback(UpperCamelCase_ ) else: __lowerCamelCase = single_urls download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = [os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(Path(UpperCamelCase_ ).name ) ) for x in single_urls] else: __lowerCamelCase = single_urls __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(Path(UpperCamelCase_ ).name ) ) __lowerCamelCase = value # make sure that values are unique if all(isinstance(UpperCamelCase_ , UpperCamelCase_ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique __lowerCamelCase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[int] ): __lowerCamelCase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one __lowerCamelCase = all(bool(re.findall("""[0-9]{3,}-of-[0-9]{3,}""" , UpperCamelCase_ ) ) for url in data_url ) __lowerCamelCase = all( url.startswith("""https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed""" ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): __lowerCamelCase = [data_url[0]] * len(UpperCamelCase_ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(single_url.split("""/""" )[-1] ) ) dummy_data_list.append(UpperCamelCase_ ) return dummy_data_list def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[Any] ): for download_callback in self.download_callbacks: download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(data_url.split("""/""" )[-1] ) ) if os.path.exists(UpperCamelCase_ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def lowerCAmelCase__ ( self: Optional[Any] ): pass def lowerCAmelCase__ ( self: List[Any] ): pass def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Dict ): def _iter_archive_members(UpperCamelCase_: Any ): # this preserves the order of the members inside the ZIP archive __lowerCamelCase = Path(self.dummy_file ).parent __lowerCamelCase = path.relative_to(UpperCamelCase_ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: __lowerCamelCase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(UpperCamelCase_ ) __lowerCamelCase = Path(UpperCamelCase_ ) __lowerCamelCase = _iter_archive_members(UpperCamelCase_ ) if self.use_local_dummy_data else path.rglob("""*""" ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith((""".""", """__""") ): yield file_path.relative_to(UpperCamelCase_ ).as_posix(), file_path.open("""rb""" ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Dict ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = [paths] for path in paths: if os.path.isfile(UpperCamelCase_ ): if os.path.basename(UpperCamelCase_ ).startswith((""".""", """__""") ): return yield path else: for dirpath, dirnames, filenames in os.walk(UpperCamelCase_ ): if os.path.basename(UpperCamelCase_ ).startswith((""".""", """__""") ): continue dirnames.sort() for filename in sorted(UpperCamelCase_ ): if filename.startswith((""".""", """__""") ): continue yield os.path.join(UpperCamelCase_ , UpperCamelCase_ )
12
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase__ = logging.get_logger(__name__) def _a ( a :str , a :Union[str, Any]=False ) -> List[Any]: a = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" a = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def _a ( a :Dict , a :Any , a :int=False ) -> Optional[Any]: for i in range(config.num_hidden_layers ): if base_model: a = '''''' else: a = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) a = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) a = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict a = in_proj_weight[ : config.hidden_size, : ] a = in_proj_bias[: config.hidden_size] a = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] a = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] a = in_proj_weight[ -config.hidden_size :, : ] a = in_proj_bias[-config.hidden_size :] def _a ( a :Tuple ) -> Dict: a = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(a , a ) def _a ( a :Optional[int] , a :Tuple , a :Union[str, Any] ) -> Any: a = dct.pop(a ) a = val def _a ( ) -> Union[str, Any]: a = '''http://images.cocodataset.org/val2017/000000039769.jpg''' a = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def _a ( a :Dict , a :str , a :Optional[int]=True ) -> Optional[int]: a = ViTConfig() # patch_size if model_name[-1] == "8": a = 8 # set labels if required if not base_model: a = 1_000 a = '''huggingface/label-files''' a = '''imagenet-1k-id2label.json''' a = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) a = {int(a ): v for k, v in idalabel.items()} a = idalabel a = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: a = 384 a = 1_536 a = 12 a = 6 # load original model from torch hub a = torch.hub.load('''facebookresearch/dino:main''' , a ) original_model.eval() # load state_dict of original model, remove and rename some keys a = original_model.state_dict() if base_model: remove_classification_head_(a ) a = create_rename_keys(a , base_model=a ) for src, dest in rename_keys: rename_key(a , a , a ) read_in_q_k_v(a , a , a ) # load HuggingFace model if base_model: a = ViTModel(a , add_pooling_layer=a ).eval() else: a = ViTForImageClassification(a ).eval() model.load_state_dict(a ) # Check outputs on an image, prepared by ViTImageProcessor a = ViTImageProcessor() a = image_processor(images=prepare_img() , return_tensors='''pt''' ) a = encoding['''pixel_values'''] a = model(a ) if base_model: a = original_model(a ) assert torch.allclose(a , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: a = original_model(a ) assert logits.shape == outputs.logits.shape assert torch.allclose(a , outputs.logits , atol=1e-3 ) Path(a ).mkdir(exist_ok=a ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="dino_vitb16", type=str, help="Name of the model trained with DINO you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--base_model", action="store_true", help="Whether to only convert the base model (no projection head weights).", ) parser.set_defaults(base_model=True) UpperCAmelCase__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
26
import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def _a ( a :Tuple ) -> int: a = tmp_path / '''file.csv''' a = textwrap.dedent( '''\ header1,header2 1,2 10,20 ''' ) with open(a , '''w''' ) as f: f.write(a ) return str(a ) @pytest.fixture def _a ( a :int ) -> List[str]: a = tmp_path / '''malformed_file.csv''' a = textwrap.dedent( '''\ header1,header2 1,2 10,20, ''' ) with open(a , '''w''' ) as f: f.write(a ) return str(a ) @pytest.fixture def _a ( a :Dict , a :int ) -> List[str]: a = tmp_path / '''csv_with_image.csv''' a = textwrap.dedent( F"""\ image {image_file} """ ) with open(a , '''w''' ) as f: f.write(a ) return str(a ) @pytest.fixture def _a ( a :List[Any] ) -> Dict: a = tmp_path / '''csv_with_label.csv''' a = textwrap.dedent( '''\ label good bad good ''' ) with open(a , '''w''' ) as f: f.write(a ) return str(a ) @pytest.fixture def _a ( a :Tuple ) -> Any: a = tmp_path / '''csv_with_int_list.csv''' a = textwrap.dedent( '''\ int_list 1 2 3 4 5 6 7 8 9 ''' ) with open(a , '''w''' ) as f: f.write(a ) return str(a ) def _a ( a :Dict , a :int , a :Union[str, Any] ) -> List[Any]: a = Csv() a = csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(a , match='''Error tokenizing data''' ): for _ in generator: pass assert any( record.levelname == '''ERROR''' and '''Failed to read file''' in record.message and os.path.basename(a ) in record.message for record in caplog.records ) @require_pil def _a ( a :Dict ) -> Any: with open(a , encoding='''utf-8''' ) as f: a = f.read().splitlines()[1] a = Csv(encoding='''utf-8''' , features=Features({'''image''': Image()} ) ) a = csv._generate_tables([[csv_file_with_image]] ) a = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''image''' ).type == Image()() a = pa_table.to_pydict()['''image'''] assert generated_content == [{"path": image_file, "bytes": None}] def _a ( a :Any ) -> Tuple: with open(a , encoding='''utf-8''' ) as f: a = f.read().splitlines()[1:] a = Csv(encoding='''utf-8''' , features=Features({'''label''': ClassLabel(names=['''good''', '''bad'''] )} ) ) a = csv._generate_tables([[csv_file_with_label]] ) a = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''label''' ).type == ClassLabel(names=['''good''', '''bad'''] )() a = pa_table.to_pydict()['''label'''] assert generated_content == [ClassLabel(names=['''good''', '''bad'''] ).straint(a ) for label in labels] def _a ( a :Union[str, Any] ) -> Optional[Any]: a = Csv(encoding='''utf-8''' , sep=''',''' , converters={'''int_list''': lambda a : [int(a ) for i in x.split()]} ) a = csv._generate_tables([[csv_file_with_int_list]] ) a = pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field('''int_list''' ).type ) a = pa_table.to_pydict()['''int_list'''] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
26
1
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): _lowercase : Dict = AutoencoderKL _lowercase : Tuple = """sample""" _lowercase : Optional[int] = 1E-2 @property def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : Union[str, Any] =4 a__ : List[str] =3 a__ : Union[str, Any] =(3_2, 3_2) a__ : int =floats_tensor((batch_size, num_channels) + sizes ).to(lowerCAmelCase__ ) return {"sample": image} @property def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' return (3, 3_2, 3_2) @property def _lowercase ( self ) -> List[Any]: '''simple docstring''' return (3, 3_2, 3_2) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : List[Any] ={ "block_out_channels": [3_2, 6_4], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } a__ : List[str] =self.dummy_input return init_dict, inputs_dict def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' pass def _lowercase ( self ) -> List[Any]: '''simple docstring''' pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ , a__ : Union[str, Any] =self.prepare_init_args_and_inputs_for_common() a__ : Tuple =self.model_class(**lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) assert not model.is_gradient_checkpointing and model.training a__ : Optional[Any] =model(**lowerCAmelCase__ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() a__ : str =torch.randn_like(lowerCAmelCase__ ) a__ : Tuple =(out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing a__ : Any =self.model_class(**lowerCAmelCase__ ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCAmelCase__ ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training a__ : Dict =model_a(**lowerCAmelCase__ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() a__ : int =(out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) a__ : Union[str, Any] =dict(model.named_parameters() ) a__ : int =dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ , a__ : Optional[int] =AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCAmelCase__ ) a__ : Optional[Any] =model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Tuple =AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) a__ : str =model.to(lowerCAmelCase__ ) model.eval() if torch_device == "mps": a__ : Any =torch.manual_seed(0 ) else: a__ : Union[str, Any] =torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) a__ : Union[str, Any] =torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) a__ : Union[str, Any] =image.to(lowerCAmelCase__ ) with torch.no_grad(): a__ : Tuple =model(lowerCAmelCase__ , sample_posterior=lowerCAmelCase__ , generator=lowerCAmelCase__ ).sample a__ : Any =output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": a__ : str =torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": a__ : Dict =torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: a__ : Optional[int] =torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1E-2 ) ) @slow class __lowerCAmelCase ( unittest.TestCase): def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' return F'''gaussian_noise_s={seed}_shape={"_".join([str(lowerCAmelCase__ ) for s in shape] )}.npy''' def _lowercase ( self ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self , lowerCAmelCase__=0 , lowerCAmelCase__=(4, 3, 5_1_2, 5_1_2) , lowerCAmelCase__=False ) -> Optional[Any]: '''simple docstring''' a__ : List[str] =torch.floataa if fpaa else torch.floataa a__ : Tuple =torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCAmelCase__ , lowerCAmelCase__ ) ) ).to(lowerCAmelCase__ ).to(lowerCAmelCase__ ) return image def _lowercase ( self , lowerCAmelCase__="CompVis/stable-diffusion-v1-4" , lowerCAmelCase__=False ) -> Any: '''simple docstring''' a__ : str ="fp16" if fpaa else None a__ : str =torch.floataa if fpaa else torch.floataa a__ : Dict =AutoencoderKL.from_pretrained( lowerCAmelCase__ , subfolder="vae" , torch_dtype=lowerCAmelCase__ , revision=lowerCAmelCase__ , ) model.to(lowerCAmelCase__ ).eval() return model def _lowercase ( self , lowerCAmelCase__=0 ) -> Optional[int]: '''simple docstring''' if torch_device == "mps": return torch.manual_seed(lowerCAmelCase__ ) return torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) @parameterized.expand( [ # fmt: off [3_3, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [4_7, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: '''simple docstring''' a__ : List[Any] =self.get_sd_vae_model() a__ : Tuple =self.get_sd_image(lowerCAmelCase__ ) a__ : Tuple =self.get_generator(lowerCAmelCase__ ) with torch.no_grad(): a__ : Tuple =model(lowerCAmelCase__ , generator=lowerCAmelCase__ , sample_posterior=lowerCAmelCase__ ).sample assert sample.shape == image.shape a__ : Tuple =sample[-1, -2:, -2:, :2].flatten().float().cpu() a__ : Tuple =torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=3E-3 ) @parameterized.expand( [ # fmt: off [3_3, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [4_7, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' a__ : Optional[int] =self.get_sd_vae_model(fpaa=lowerCAmelCase__ ) a__ : Any =self.get_sd_image(lowerCAmelCase__ , fpaa=lowerCAmelCase__ ) a__ : Dict =self.get_generator(lowerCAmelCase__ ) with torch.no_grad(): a__ : Union[str, Any] =model(lowerCAmelCase__ , generator=lowerCAmelCase__ , sample_posterior=lowerCAmelCase__ ).sample assert sample.shape == image.shape a__ : int =sample[-1, -2:, :2, -2:].flatten().float().cpu() a__ : List[str] =torch.tensor(lowerCAmelCase__ ) assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [4_7, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: '''simple docstring''' a__ : Any =self.get_sd_vae_model() a__ : Optional[int] =self.get_sd_image(lowerCAmelCase__ ) with torch.no_grad(): a__ : Dict =model(lowerCAmelCase__ ).sample assert sample.shape == image.shape a__ : List[str] =sample[-1, -2:, -2:, :2].flatten().float().cpu() a__ : List[str] =torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=3E-3 ) @parameterized.expand( [ # fmt: off [1_3, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [3_7, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' a__ : List[Any] =self.get_sd_vae_model() a__ : Optional[Any] =self.get_sd_image(lowerCAmelCase__ , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): a__ : Optional[Any] =model.decode(lowerCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] a__ : List[Any] =sample[-1, -2:, :2, -2:].flatten().cpu() a__ : Dict =torch.tensor(lowerCAmelCase__ ) assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) @parameterized.expand( [ # fmt: off [2_7, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [1_6, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' a__ : Optional[Any] =self.get_sd_vae_model(fpaa=lowerCAmelCase__ ) a__ : Tuple =self.get_sd_image(lowerCAmelCase__ , shape=(3, 4, 6_4, 6_4) , fpaa=lowerCAmelCase__ ) with torch.no_grad(): a__ : Union[str, Any] =model.decode(lowerCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] a__ : Tuple =sample[-1, -2:, :2, -2:].flatten().float().cpu() a__ : Dict =torch.tensor(lowerCAmelCase__ ) assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=5E-3 ) @parameterized.expand([(1_3,), (1_6,), (2_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def _lowercase ( self , lowerCAmelCase__ ) -> Dict: '''simple docstring''' a__ : int =self.get_sd_vae_model(fpaa=lowerCAmelCase__ ) a__ : List[Any] =self.get_sd_image(lowerCAmelCase__ , shape=(3, 4, 6_4, 6_4) , fpaa=lowerCAmelCase__ ) with torch.no_grad(): a__ : Optional[int] =model.decode(lowerCAmelCase__ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): a__ : Tuple =model.decode(lowerCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-1 ) @parameterized.expand([(1_3,), (1_6,), (3_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def _lowercase ( self , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' a__ : Dict =self.get_sd_vae_model() a__ : str =self.get_sd_image(lowerCAmelCase__ , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): a__ : List[str] =model.decode(lowerCAmelCase__ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): a__ : Optional[int] =model.decode(lowerCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [4_7, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : Tuple =self.get_sd_vae_model() a__ : int =self.get_sd_image(lowerCAmelCase__ ) a__ : Any =self.get_generator(lowerCAmelCase__ ) with torch.no_grad(): a__ : str =model.encode(lowerCAmelCase__ ).latent_dist a__ : Union[str, Any] =dist.sample(generator=lowerCAmelCase__ ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] a__ : Optional[Any] =sample[0, -1, -3:, -3:].flatten().cpu() a__ : List[str] =torch.tensor(lowerCAmelCase__ ) a__ : Union[str, Any] =3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCAmelCase__ , lowerCAmelCase__ , atol=lowerCAmelCase__ )
95
from math import pi def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): """simple docstring""" return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
95
1
"""simple docstring""" from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class _a (snake_case_ ): '''simple docstring''' def __init__( self , A__ , A__ = None , A__ = None , A__ = True , A__ = None , A__ = False , A__ = None , A__ = True , A__ = "arrow" , **A__ , ): super().__init__( split=A__ , features=A__ , cache_dir=A__ , keep_in_memory=A__ , streaming=A__ , **A__ , ) A__ : Union[str, Any] = load_from_cache_file A__ : Tuple = file_format A__ : Any = Spark( df=A__ , features=A__ , cache_dir=A__ , working_dir=A__ , **A__ , ) def __A ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) A__ : Union[str, Any] = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=A__ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
353
from queue import PriorityQueue from typing import Any import numpy as np def UpperCamelCase (lowercase_: dict , lowercase_: str , lowercase_: set , lowercase_: set , lowercase_: dict , lowercase_: dict , lowercase_: PriorityQueue , lowercase_: dict , lowercase_: float | int , ) -> float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue A__ : Any = cst_fwd.get(lowercase_ , np.inf ) A__ : List[Any] = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) A__ : Tuple = new_cost_f A__ : Any = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: A__ : Optional[int] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def UpperCamelCase (lowercase_: str , lowercase_: str , lowercase_: dict , lowercase_: dict ) -> int: A__ : Dict = -1 A__ : List[Any] = set() A__ : Union[str, Any] = set() A__ : Optional[Any] = {source: 0} A__ : int = {destination: 0} A__ : Optional[Any] = {source: None} A__ : Union[str, Any] = {destination: None} A__ : PriorityQueue[Any] = PriorityQueue() A__ : PriorityQueue[Any] = PriorityQueue() A__ : List[Any] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): A__ , A__ : Tuple = queue_forward.get() visited_forward.add(lowercase_ ) A__ , A__ : Optional[Any] = queue_backward.get() visited_backward.add(lowercase_ ) A__ : List[Any] = pass_and_relaxation( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) A__ : List[Any] = pass_and_relaxation( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: A__ : int = shortest_distance return shortest_path_distance A_ : List[Any] = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } A_ : Optional[int] = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
141
0
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) lowercase : List[Any] = { "iou_prediction_head.layers.0": "iou_prediction_head.proj_in", "iou_prediction_head.layers.1": "iou_prediction_head.layers.0", "iou_prediction_head.layers.2": "iou_prediction_head.proj_out", "mask_decoder.output_upscaling.0": "mask_decoder.upscale_conv1", "mask_decoder.output_upscaling.1": "mask_decoder.upscale_layer_norm", "mask_decoder.output_upscaling.3": "mask_decoder.upscale_conv2", "mask_downscaling.0": "mask_embed.conv1", "mask_downscaling.1": "mask_embed.layer_norm1", "mask_downscaling.3": "mask_embed.conv2", "mask_downscaling.4": "mask_embed.layer_norm2", "mask_downscaling.6": "mask_embed.conv3", "point_embeddings": "point_embed", "pe_layer.positional_encoding_gaussian_matrix": "shared_embedding.positional_embedding", "image_encoder": "vision_encoder", "neck.0": "neck.conv1", "neck.1": "neck.layer_norm1", "neck.2": "neck.conv2", "neck.3": "neck.layer_norm2", "patch_embed.proj": "patch_embed.projection", ".norm": ".layer_norm", "blocks": "layers", } def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : List[str]) -> Any: '''simple docstring''' __UpperCamelCase : List[Any] = {} state_dict.pop("pixel_mean" , __lowerCAmelCase) state_dict.pop("pixel_std" , __lowerCAmelCase) __UpperCamelCase : List[str] = R'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __UpperCamelCase : Optional[int] = key.replace(__lowerCAmelCase , __lowerCAmelCase) if re.match(__lowerCAmelCase , __lowerCAmelCase): __UpperCamelCase : Union[str, Any] = int(re.match(__lowerCAmelCase , __lowerCAmelCase).group(2)) if layer_nb == 0: __UpperCamelCase : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: __UpperCamelCase : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: __UpperCamelCase : Optional[int] = key.replace("layers.2" , "proj_out") __UpperCamelCase : str = value __UpperCamelCase : List[Any] = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Dict , _lowerCamelCase : Dict , _lowerCamelCase : str , _lowerCamelCase : Any="ybelkada/segment-anything") -> Optional[int]: '''simple docstring''' __UpperCamelCase : Optional[Any] = hf_hub_download(__lowerCAmelCase , F'checkpoints/{model_name}.pth') if "sam_vit_b" in model_name: __UpperCamelCase : Tuple = SamConfig() elif "sam_vit_l" in model_name: __UpperCamelCase : List[str] = SamVisionConfig( hidden_size=1_024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) __UpperCamelCase : int = SamConfig( vision_config=__lowerCAmelCase , ) elif "sam_vit_h" in model_name: __UpperCamelCase : Union[str, Any] = SamVisionConfig( hidden_size=1_280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) __UpperCamelCase : str = SamConfig( vision_config=__lowerCAmelCase , ) __UpperCamelCase : List[str] = torch.load(__lowerCAmelCase , map_location="cpu") __UpperCamelCase : Dict = replace_keys(__lowerCAmelCase) __UpperCamelCase : int = SamImageProcessor() __UpperCamelCase : Any = SamProcessor(image_processor=__lowerCAmelCase) __UpperCamelCase : Dict = SamModel(__lowerCAmelCase) hf_model.load_state_dict(__lowerCAmelCase) __UpperCamelCase : Optional[int] = hf_model.to("cuda") __UpperCamelCase : List[Any] = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' __UpperCamelCase : int = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase).raw).convert("RGB") __UpperCamelCase : List[str] = [[[400, 650]]] __UpperCamelCase : List[Any] = [[1]] __UpperCamelCase : Optional[int] = processor(images=np.array(__lowerCAmelCase) , return_tensors="pt").to("cuda") with torch.no_grad(): __UpperCamelCase : Tuple = hf_model(**__lowerCAmelCase) __UpperCamelCase : List[Any] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.5_7_9_8_9_0_2_5_1_1_5_9_6_6_8 __UpperCamelCase : Any = processor( images=np.array(__lowerCAmelCase) , input_points=__lowerCAmelCase , input_labels=__lowerCAmelCase , return_tensors="pt").to("cuda") with torch.no_grad(): __UpperCamelCase : int = hf_model(**__lowerCAmelCase) __UpperCamelCase : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_7_1_2_6_0_3_0_9_2_1_9_3_6_0_4 __UpperCamelCase : str = ((75, 275, 1_725, 850),) __UpperCamelCase : Optional[int] = processor(images=np.array(__lowerCAmelCase) , input_boxes=__lowerCAmelCase , return_tensors="pt").to("cuda") with torch.no_grad(): __UpperCamelCase : Optional[int] = hf_model(**__lowerCAmelCase) __UpperCamelCase : Optional[Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_6_8_6_0_1_5_6_0_5_9_2_6_5_1_4 # Test with 2 points and 1 image. __UpperCamelCase : List[Any] = [[[400, 650], [800, 650]]] __UpperCamelCase : List[Any] = [[1, 1]] __UpperCamelCase : Any = processor( images=np.array(__lowerCAmelCase) , input_points=__lowerCAmelCase , input_labels=__lowerCAmelCase , return_tensors="pt").to("cuda") with torch.no_grad(): __UpperCamelCase : str = hf_model(**__lowerCAmelCase) __UpperCamelCase : List[str] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_9_3_6_0_4_7_7_9_2_4_3_4_6_9_2 if __name__ == "__main__": lowercase : Tuple = argparse.ArgumentParser() lowercase : List[str] = ["sam_vit_b_01ec64", "sam_vit_h_4b8939", "sam_vit_l_0b3195"] parser.add_argument( '--model_name', default='sam_vit_h_4b8939', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) parser.add_argument( '--model_hub_id', default='ybelkada/segment-anything', choices=choices, type=str, help='Path to hf config.json of model to convert', ) lowercase : Optional[Any] = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
232
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass __lowerCAmelCase : Any = (3, 9, -11, 0, 7, 5, 1, -1) __lowerCAmelCase : Tuple = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class __lowerCAmelCase : """simple docstring""" A__ : int A__ : Node | None class __lowerCAmelCase : """simple docstring""" def __init__( self : Optional[Any] , _snake_case : Iterable[int] ): __lowercase : Node | None = None for i in sorted(_snake_case , reverse=_snake_case ): __lowercase : List[Any] = Node(_snake_case , self.head ) def __iter__( self : str ): __lowercase : Union[str, Any] = self.head while node: yield node.data __lowercase : List[Any] = node.next_node def __len__( self : str ): return sum(1 for _ in self ) def __str__( self : List[str] ): return " -> ".join([str(_snake_case ) for node in self] ) def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase ) -> SortedLinkedList: return SortedLinkedList(list(__lowerCAmelCase ) + list(__lowerCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase : Dict = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
156
0
class UpperCamelCase_ : '''simple docstring''' def __init__( self : List[str] , UpperCAmelCase__ : int) ->Union[str, Any]: '''simple docstring''' A__ = n A__ = [None] * self.n A__ = 0 # index of the first element A__ = 0 A__ = 0 def __len__( self : List[Any]) ->int: '''simple docstring''' return self.size def SCREAMING_SNAKE_CASE ( self : List[Any]) ->bool: '''simple docstring''' return self.size == 0 def SCREAMING_SNAKE_CASE ( self : str) ->List[Any]: '''simple docstring''' return False if self.is_empty() else self.array[self.front] def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : str) ->Any: '''simple docstring''' if self.size >= self.n: raise Exception('''QUEUE IS FULL''') A__ = data A__ = (self.rear + 1) % self.n self.size += 1 return self def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Any: '''simple docstring''' if self.size == 0: raise Exception('''UNDERFLOW''') A__ = self.array[self.front] A__ = None A__ = (self.front + 1) % self.n self.size -= 1 return temp
231
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCamelCase : Any = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : str = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
231
1
'''simple docstring''' import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''', [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 10, '''max_num_jobs''': 1}, [range(10 )]), ({'''num_shards''': 10, '''max_num_jobs''': 10}, [range(lowerCAmelCase__, i + 1 ) for i in range(10 )]), ({'''num_shards''': 1, '''max_num_jobs''': 10}, [range(1 )]), ({'''num_shards''': 10, '''max_num_jobs''': 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({'''num_shards''': 3, '''max_num_jobs''': 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def a__ ( lowercase : List[Any], lowercase : Optional[int] ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = _distribute_shards(**lowerCAmelCase__ ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''', [ ({'''foo''': 0}, 10, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ], ) def a__ ( lowercase : List[Any], lowercase : str, lowercase : str ) -> Any: """simple docstring""" _UpperCamelCase = _split_gen_kwargs(lowerCAmelCase__, lowerCAmelCase__ ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''', [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ], ) def a__ ( lowercase : Any, lowercase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" if expected is RuntimeError: with pytest.raises(lowerCAmelCase__ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase__ ) else: _UpperCamelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase__ ) assert out == expected
324
from __future__ import annotations from typing import Any class _a : def __init__(self, SCREAMING_SNAKE_CASE_ = 6 ) -> None: UpperCAmelCase_: Node | None = None UpperCAmelCase_: Node | None = None self.create_linked_list(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCAmelCase_: Optional[Any] = Node() UpperCAmelCase_: Optional[Any] = current_node UpperCAmelCase_: List[str] = current_node UpperCAmelCase_: List[Any] = current_node for _ in range(1, SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Optional[int] = Node() UpperCAmelCase_: Dict = current_node UpperCAmelCase_: Any = previous_node UpperCAmelCase_: Tuple = current_node UpperCAmelCase_: Optional[Any] = self.front UpperCAmelCase_: Any = previous_node def __snake_case (self ) -> bool: return ( self.front == self.rear and self.front is not None and self.front.data is None ) def __snake_case (self ) -> Any | None: self.check_can_perform_operation() return self.front.data if self.front else None def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> None: if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase_: Optional[int] = self.rear.next if self.rear: UpperCAmelCase_: Any = data def __snake_case (self ) -> Any: self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase_: Union[str, Any] = self.front.data UpperCAmelCase_: Any = None return data UpperCAmelCase_: str = self.front UpperCAmelCase_: Union[str, Any] = old_front.next UpperCAmelCase_: int = old_front.data UpperCAmelCase_: Any = None return data def __snake_case (self ) -> None: if self.is_empty(): raise Exception("""Empty Queue""" ) def __snake_case (self ) -> None: if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class _a : def __init__(self ) -> None: UpperCAmelCase_: Any | None = None UpperCAmelCase_: Node | None = None UpperCAmelCase_: Node | None = None if __name__ == "__main__": import doctest doctest.testmod()
147
0
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase = 1000 ): lowercase__ : Dict = 2**power lowercase__ : List[Any] = 0 while n: lowercase__ , lowercase__ : List[str] = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
214
'''simple docstring''' class UpperCAmelCase : '''simple docstring''' def __init__( self ) -> List[str]: lowercase__ : Dict = {} def _lowerCAmelCase( self ) -> None: print(self.vertex ) for i in self.vertex: print(__lowerCAmelCase , ''' -> ''' , ''' -> '''.join([str(__lowerCAmelCase ) for j in self.vertex[i]] ) ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(__lowerCAmelCase ) else: # else make a new vertex lowercase__ : Union[str, Any] = [to_vertex] def _lowerCAmelCase( self ) -> None: # visited array for storing already visited nodes lowercase__ : str = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> None: # mark start vertex as visited lowercase__ : List[str] = True print(__lowerCAmelCase , end=''' ''' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": __a: Optional[Any] = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
214
1
lowerCamelCase : Any =[4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCamelCase : str =[3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCamelCase : Dict ={ 0: '''Sunday''', 1: '''Monday''', 2: '''Tuesday''', 3: '''Wednesday''', 4: '''Thursday''', 5: '''Friday''', 6: '''Saturday''', } def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: assert len(str(__lowerCAmelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: UpperCamelCase__ : List[str] = year // 100 UpperCamelCase__ : str = (5 * (century % 4) + 2) % 7 UpperCamelCase__ : int = year % 100 UpperCamelCase__ : Optional[Any] = centurian % 12 UpperCamelCase__ : int = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 UpperCamelCase__ : str = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) UpperCamelCase__ : Optional[int] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
189
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase : Tuple ={ '''configuration_luke''': ['''LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LukeConfig'''], '''tokenization_luke''': ['''LukeTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Union[str, Any] =[ '''LUKE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LukeForEntityClassification''', '''LukeForEntityPairClassification''', '''LukeForEntitySpanClassification''', '''LukeForMultipleChoice''', '''LukeForQuestionAnswering''', '''LukeForSequenceClassification''', '''LukeForTokenClassification''', '''LukeForMaskedLM''', '''LukeModel''', '''LukePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase : Dict =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
189
1
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
368
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case__ : List[Any] = logging.get_logger(__name__) snake_case__ : Optional[Any] = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ ): '''simple docstring''' lowerCamelCase_ :List[str] = '''autoformer''' lowerCamelCase_ :Optional[Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self , snake_case_ = None , snake_case_ = None , snake_case_ = "student_t" , snake_case_ = "nll" , snake_case_ = 1 , snake_case_ = [1, 2, 3, 4, 5, 6, 7] , snake_case_ = True , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = 6_4 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 3_2 , snake_case_ = 3_2 , snake_case_ = "gelu" , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 1_0_0 , snake_case_ = 0.02 , snake_case_ = True , snake_case_=True , snake_case_ = 1_0 , snake_case_ = 2_5 , snake_case_ = 3 , **snake_case_ , ): '''simple docstring''' UpperCAmelCase_ : List[Any] = prediction_length UpperCAmelCase_ : List[str] = context_length if context_length is not None else prediction_length UpperCAmelCase_ : Optional[int] = distribution_output UpperCAmelCase_ : Optional[int] = loss UpperCAmelCase_ : Union[str, Any] = input_size UpperCAmelCase_ : int = num_time_features UpperCAmelCase_ : List[str] = lags_sequence UpperCAmelCase_ : Any = scaling UpperCAmelCase_ : Any = num_dynamic_real_features UpperCAmelCase_ : int = num_static_real_features UpperCAmelCase_ : Optional[Any] = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) UpperCAmelCase_ : List[Any] = cardinality else: UpperCAmelCase_ : Optional[int] = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) UpperCAmelCase_ : List[str] = embedding_dimension else: UpperCAmelCase_ : List[Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase_ : List[str] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase_ : Union[str, Any] = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase_ : str = d_model UpperCAmelCase_ : str = encoder_attention_heads UpperCAmelCase_ : str = decoder_attention_heads UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : str = decoder_ffn_dim UpperCAmelCase_ : str = encoder_layers UpperCAmelCase_ : str = decoder_layers UpperCAmelCase_ : str = dropout UpperCAmelCase_ : Optional[int] = attention_dropout UpperCAmelCase_ : Tuple = activation_dropout UpperCAmelCase_ : Any = encoder_layerdrop UpperCAmelCase_ : Tuple = decoder_layerdrop UpperCAmelCase_ : List[str] = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Union[str, Any] = use_cache # Autoformer UpperCAmelCase_ : Any = label_length UpperCAmelCase_ : Union[str, Any] = moving_average UpperCAmelCase_ : Tuple = autocorrelation_factor super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def _UpperCamelCase ( self ): '''simple docstring''' return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
274
0
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowerCamelCase : Tuple = logging.get_logger(__name__) _lowerCamelCase : Tuple = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } _lowerCamelCase : Optional[Any] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } _lowerCamelCase : str = { """facebook/blenderbot_small-90M""": 512, } class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = BlenderbotSmallTokenizer def __init__( self : List[str] , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Optional[Any]="<|endoftext|>" , UpperCAmelCase__ : Union[str, Any]="<|endoftext|>" , UpperCAmelCase__ : Any="<|endoftext|>" , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : int=True , **UpperCAmelCase__ : Dict , ) ->Tuple: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=UpperCAmelCase__ , merges=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , trim_offsets=UpperCAmelCase__ , ) , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) A__ = add_prefix_space def SCREAMING_SNAKE_CASE ( self : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any]=None) ->Dict: '''simple docstring''' A__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None) ->List[int]: '''simple docstring''' A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
14
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> List[Any]: """simple docstring""" A__ = StableDiffusionPipeline.from_pretrained(lowercase_ , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors A__ = load_file(lowercase_ ) A__ = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: A__ = key.split('''.''' )[0].split(LORA_PREFIX_TEXT_ENCODER + '''_''' )[-1].split('''_''' ) A__ = pipeline.text_encoder else: A__ = key.split('''.''' )[0].split(LORA_PREFIX_UNET + '''_''' )[-1].split('''_''' ) A__ = pipeline.unet # find the target layer A__ = layer_infos.pop(0 ) while len(lowercase_ ) > -1: try: A__ = curr_layer.__getattr__(lowercase_ ) if len(lowercase_ ) > 0: A__ = layer_infos.pop(0 ) elif len(lowercase_ ) == 0: break except Exception: if len(lowercase_ ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: A__ = layer_infos.pop(0 ) A__ = [] if "lora_down" in key: pair_keys.append(key.replace('''lora_down''' , '''lora_up''' ) ) pair_keys.append(lowercase_ ) else: pair_keys.append(lowercase_ ) pair_keys.append(key.replace('''lora_up''' , '''lora_down''' ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: A__ = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) A__ = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(lowercase_ , lowercase_ ).unsqueeze(2 ).unsqueeze(3 ) else: A__ = state_dict[pair_keys[0]].to(torch.floataa ) A__ = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(lowercase_ , lowercase_ ) # update visited list for item in pair_keys: visited.append(lowercase_ ) return pipeline if __name__ == "__main__": _lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--base_model_path""", default=None, type=str, required=True, help="""Path to the base model in diffusers format.""" ) parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--lora_prefix_unet""", default="""lora_unet""", type=str, help="""The prefix of UNet weight in safetensors""" ) parser.add_argument( """--lora_prefix_text_encoder""", default="""lora_te""", type=str, help="""The prefix of text encoder weight in safetensors""", ) parser.add_argument("""--alpha""", default=0.75, type=float, help="""The merging ratio in W = W0 + alpha * deltaW""") parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""" ) parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") _lowerCamelCase : Tuple = parser.parse_args() _lowerCamelCase : List[Any] = args.base_model_path _lowerCamelCase : Optional[int] = args.checkpoint_path _lowerCamelCase : Dict = args.dump_path _lowerCamelCase : Optional[Any] = args.lora_prefix_unet _lowerCamelCase : Optional[int] = args.lora_prefix_text_encoder _lowerCamelCase : List[Any] = args.alpha _lowerCamelCase : int = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) _lowerCamelCase : Tuple = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
14
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer A_ : Tuple = logging.get_logger(__name__) A_ : Tuple = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} A_ : List[str] = { 'vocab_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/vocab.txt', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/vocab.txt', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt' ), 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt' ), 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt', 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json' ), 'bert-base-multilingual-cased': ( 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json' ), 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-cased': ( 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json' ), }, } A_ : Dict = { 'bert-base-uncased': 512, 'bert-large-uncased': 512, 'bert-base-cased': 512, 'bert-large-cased': 512, 'bert-base-multilingual-uncased': 512, 'bert-base-multilingual-cased': 512, 'bert-base-chinese': 512, 'bert-base-german-cased': 512, 'bert-large-uncased-whole-word-masking': 512, 'bert-large-cased-whole-word-masking': 512, 'bert-large-uncased-whole-word-masking-finetuned-squad': 512, 'bert-large-cased-whole-word-masking-finetuned-squad': 512, 'bert-base-cased-finetuned-mrpc': 512, 'bert-base-german-dbmdz-cased': 512, 'bert-base-german-dbmdz-uncased': 512, 'TurkuNLP/bert-base-finnish-cased-v1': 512, 'TurkuNLP/bert-base-finnish-uncased-v1': 512, 'wietsedv/bert-base-dutch-cased': 512, } A_ : Any = { 'bert-base-uncased': {'do_lower_case': True}, 'bert-large-uncased': {'do_lower_case': True}, 'bert-base-cased': {'do_lower_case': False}, 'bert-large-cased': {'do_lower_case': False}, 'bert-base-multilingual-uncased': {'do_lower_case': True}, 'bert-base-multilingual-cased': {'do_lower_case': False}, 'bert-base-chinese': {'do_lower_case': False}, 'bert-base-german-cased': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking-finetuned-squad': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking-finetuned-squad': {'do_lower_case': False}, 'bert-base-cased-finetuned-mrpc': {'do_lower_case': False}, 'bert-base-german-dbmdz-cased': {'do_lower_case': False}, 'bert-base-german-dbmdz-uncased': {'do_lower_case': True}, 'TurkuNLP/bert-base-finnish-cased-v1': {'do_lower_case': False}, 'TurkuNLP/bert-base-finnish-uncased-v1': {'do_lower_case': True}, 'wietsedv/bert-base-dutch-cased': {'do_lower_case': False}, } class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Optional[int] = VOCAB_FILES_NAMES UpperCAmelCase__: Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__: int = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__: Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__: Any = BertTokenizer def __init__( self , A__=None , A__=None , A__=True , A__="[UNK]" , A__="[SEP]" , A__="[PAD]" , A__="[CLS]" , A__="[MASK]" , A__=True , A__=None , **A__ , ): super().__init__( A__ , tokenizer_file=A__ , do_lower_case=A__ , unk_token=A__ , sep_token=A__ , pad_token=A__ , cls_token=A__ , mask_token=A__ , tokenize_chinese_chars=A__ , strip_accents=A__ , **A__ , ) A__ : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , A__ ) != do_lower_case or normalizer_state.get("""strip_accents""" , A__ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , A__ ) != tokenize_chinese_chars ): A__ : List[Any] = getattr(A__ , normalizer_state.pop("""type""" ) ) A__ : Dict = do_lower_case A__ : Optional[int] = strip_accents A__ : int = tokenize_chinese_chars A__ : Dict = normalizer_class(**A__ ) A__ : Optional[int] = do_lower_case def __A ( self , A__ , A__=None ): A__ : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __A ( self , A__ , A__ = None ): A__ : str = [self.sep_token_id] A__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __A ( self , A__ , A__ = None ): A__ : Union[str, Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ )
141
from __future__ import annotations def UpperCamelCase (lowercase_: float , lowercase_: float , lowercase_: float ) -> dict[str, float]: if (voltage, current, resistance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if resistance < 0: raise ValueError("""Resistance cannot be negative""" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
141
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase__ :Union[str, Any] = {"configuration_unispeech": ["UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP", "UniSpeechConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :str = [ "UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST", "UniSpeechForCTC", "UniSpeechForPreTraining", "UniSpeechForSequenceClassification", "UniSpeechModel", "UniSpeechPreTrainedModel", ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys lowercase__ :Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
101
'''simple docstring''' from __future__ import annotations _A : Any ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class _lowercase : def __init__( self: Tuple , UpperCamelCase__: dict[str, list[str]] , UpperCamelCase__: str ): lowerCamelCase__ : str = graph # mapping node to its parent in resulting breadth first tree lowerCamelCase__ : dict[str, str | None] = {} lowerCamelCase__ : Any = source_vertex def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = {self.source_vertex} lowerCamelCase__ : List[str] = None lowerCamelCase__ : Tuple = [self.source_vertex] # first in first out queue while queue: lowerCamelCase__ : Tuple = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase__ ) lowerCamelCase__ : List[str] = vertex queue.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): if target_vertex == self.source_vertex: return self.source_vertex lowerCamelCase__ : Tuple = self.parent.get(UpperCamelCase__ ) if target_vertex_parent is None: lowerCamelCase__ : int = ( F'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(UpperCamelCase__ ) return self.shortest_path(UpperCamelCase__ ) + F'''->{target_vertex}''' if __name__ == "__main__": _A : int =Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
41
0
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position __A = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("3.7"): raise ImportWarning( "To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition." ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( "To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n" "If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`." ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip __A = concatenate_datasets __A = DownloadConfig __A = DownloadManager __A = DownloadMode __A = DownloadConfig __A = DownloadMode __A = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
361
import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = ["model.decoder.embed_positions.weights"] def lowerCamelCase_ ( UpperCamelCase__ : List[Any] ) -> List[Any]: """simple docstring""" if "emb" in name: __lowerCamelCase = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: __lowerCamelCase = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: __lowerCamelCase = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: __lowerCamelCase = name.replace('linear1' , 'fc1' ) if "linear2" in name: __lowerCamelCase = name.replace('linear2' , 'fc2' ) if "norm1" in name: __lowerCamelCase = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: __lowerCamelCase = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: __lowerCamelCase = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: __lowerCamelCase = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: __lowerCamelCase = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: __lowerCamelCase = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def lowerCamelCase_ ( UpperCamelCase__ : OrderedDict , UpperCamelCase__ : int ) -> Tuple[Dict, Dict]: """simple docstring""" __lowerCamelCase = list(state_dict.keys() ) __lowerCamelCase = {} for key in keys: __lowerCamelCase = state_dict.pop(UpperCamelCase__ ) __lowerCamelCase = rename_keys(UpperCamelCase__ ) if "in_proj_weight" in key: # split fused qkv proj __lowerCamelCase = val[:hidden_size, :] __lowerCamelCase = val[hidden_size : 2 * hidden_size, :] __lowerCamelCase = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: __lowerCamelCase = val else: __lowerCamelCase = val return state_dict, enc_dec_proj_state_dict def lowerCamelCase_ ( UpperCamelCase__ : str ) -> MusicgenDecoderConfig: """simple docstring""" if checkpoint == "small": # default config values __lowerCamelCase = 1024 __lowerCamelCase = 24 __lowerCamelCase = 16 elif checkpoint == "medium": __lowerCamelCase = 1536 __lowerCamelCase = 48 __lowerCamelCase = 24 elif checkpoint == "large": __lowerCamelCase = 2048 __lowerCamelCase = 48 __lowerCamelCase = 32 else: raise ValueError(F"""Checkpoint should be one of `['small', 'medium', 'large']`, got {checkpoint}.""" ) __lowerCamelCase = MusicgenDecoderConfig( hidden_size=UpperCamelCase__ , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase__ , num_attention_heads=UpperCamelCase__ , ) return config @torch.no_grad() def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Dict=None , UpperCamelCase__ : Dict=None , UpperCamelCase__ : Optional[int]="cpu" ) -> List[Any]: """simple docstring""" __lowerCamelCase = MusicGen.get_pretrained(UpperCamelCase__ , device=UpperCamelCase__ ) __lowerCamelCase = decoder_config_from_checkpoint(UpperCamelCase__ ) __lowerCamelCase = fairseq_model.lm.state_dict() __lowerCamelCase , __lowerCamelCase = rename_state_dict( UpperCamelCase__ , hidden_size=decoder_config.hidden_size ) __lowerCamelCase = TaEncoderModel.from_pretrained('t5-base' ) __lowerCamelCase = EncodecModel.from_pretrained('facebook/encodec_32khz' ) __lowerCamelCase = MusicgenForCausalLM(UpperCamelCase__ ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection __lowerCamelCase , __lowerCamelCase = decoder.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: raise ValueError(F"""Missing key(s) in state_dict: {missing_keys}""" ) if len(UpperCamelCase__ ) > 0: raise ValueError(F"""Unexpected key(s) in state_dict: {unexpected_keys}""" ) # init the composite model __lowerCamelCase = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase__ , audio_encoder=UpperCamelCase__ , decoder=UpperCamelCase__ ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase__ ) # check we can do a forward pass __lowerCamelCase = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) __lowerCamelCase = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): __lowerCamelCase = model(input_ids=UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ ).logits if logits.shape != (8, 1, 2048): raise ValueError('Incorrect shape for logits' ) # now construct the processor __lowerCamelCase = AutoTokenizer.from_pretrained('t5-base' ) __lowerCamelCase = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) __lowerCamelCase = MusicgenProcessor(feature_extractor=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) # set the appropriate bos/pad token ids __lowerCamelCase = 2048 __lowerCamelCase = 2048 # set other default generation config params __lowerCamelCase = int(30 * audio_encoder.config.frame_rate ) __lowerCamelCase = True __lowerCamelCase = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) logger.info(F"""Saving model {checkpoint} to {pytorch_dump_folder}""" ) model.save_pretrained(UpperCamelCase__ ) processor.save_pretrained(UpperCamelCase__ ) if repo_id: logger.info(F"""Pushing model {checkpoint} to {repo_id}""" ) model.push_to_hub(UpperCamelCase__ ) processor.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint", default="small", type=str, help="Checkpoint size of the MusicGen model you'd like to convert. Can be one of: `['small', 'medium', 'large']`.", ) parser.add_argument( "--pytorch_dump_folder", required=True, default=None, type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) parser.add_argument( "--device", default="cpu", type=str, help="Torch device to run the conversion, either cpu or cuda." ) __A = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
348
0
"""simple docstring""" from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
221
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" A__ = b.T A__ = np.sum(np.square(UpperCamelCase__ ) , axis=1 ) A__ = np.sum(np.square(UpperCamelCase__ ) , axis=0 ) A__ = np.matmul(UpperCamelCase__ , UpperCamelCase__ ) A__ = aa[:, None] - 2 * ab + ba[None, :] return d def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" A__ = x.reshape(-1 , 3 ) A__ = squared_euclidean_distance(UpperCamelCase__ , UpperCamelCase__ ) return np.argmin(UpperCamelCase__ , axis=1 ) class UpperCamelCase__( __A ): lowerCAmelCase__ : Tuple = ['pixel_values'] def __init__( self ,__UpperCAmelCase = None ,__UpperCAmelCase = True ,__UpperCAmelCase = None ,__UpperCAmelCase = PILImageResampling.BILINEAR ,__UpperCAmelCase = True ,__UpperCAmelCase = True ,**__UpperCAmelCase ,) -> None: super().__init__(**__UpperCAmelCase ) A__ = size if size is not None else {'height': 2_56, 'width': 2_56} A__ = get_size_dict(__UpperCAmelCase ) A__ = np.array(__UpperCAmelCase ) if clusters is not None else None A__ = do_resize A__ = size A__ = resample A__ = do_normalize A__ = do_color_quantize def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase = PILImageResampling.BILINEAR ,__UpperCAmelCase = None ,**__UpperCAmelCase ,) -> np.ndarray: A__ = get_size_dict(__UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size dictionary must contain both height and width keys. Got {size.keys()}''' ) return resize( __UpperCAmelCase ,size=(size['height'], size['width']) ,resample=__UpperCAmelCase ,data_format=__UpperCAmelCase ,**__UpperCAmelCase ) def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ,) -> np.ndarray: A__ = rescale(image=__UpperCAmelCase ,scale=1 / 1_2_7.5 ,data_format=__UpperCAmelCase ) A__ = image - 1 return image def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase = ChannelDimension.FIRST ,**__UpperCAmelCase ,) -> PIL.Image.Image: A__ = do_resize if do_resize is not None else self.do_resize A__ = size if size is not None else self.size A__ = get_size_dict(__UpperCAmelCase ) A__ = resample if resample is not None else self.resample A__ = do_normalize if do_normalize is not None else self.do_normalize A__ = do_color_quantize if do_color_quantize is not None else self.do_color_quantize A__ = clusters if clusters is not None else self.clusters A__ = np.array(__UpperCAmelCase ) A__ = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_color_quantize and clusters is None: raise ValueError('Clusters must be specified if do_color_quantize is True.' ) # All transformations expect numpy arrays. A__ = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_resize: A__ = [self.resize(image=__UpperCAmelCase ,size=__UpperCAmelCase ,resample=__UpperCAmelCase ) for image in images] if do_normalize: A__ = [self.normalize(image=__UpperCAmelCase ) for image in images] if do_color_quantize: A__ = [to_channel_dimension_format(__UpperCAmelCase ,ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) A__ = np.array(__UpperCAmelCase ) A__ = color_quantize(__UpperCAmelCase ,__UpperCAmelCase ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) A__ = images.shape[0] A__ = images.reshape(__UpperCAmelCase ,-1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. A__ = list(__UpperCAmelCase ) else: A__ = [to_channel_dimension_format(__UpperCAmelCase ,__UpperCAmelCase ) for image in images] A__ = {'input_ids': images} return BatchFeature(data=__UpperCAmelCase ,tensor_type=__UpperCAmelCase )
221
1
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( snake_case : int , snake_case : int ) -> Any: """simple docstring""" if b == 0: return 1 if (b % 2) == 0: return actual_power(snake_case , int(b / 2 ) ) * actual_power(snake_case , int(b / 2 ) ) else: return a * actual_power(snake_case , int(b / 2 ) ) * actual_power(snake_case , int(b / 2 ) ) def SCREAMING_SNAKE_CASE__ ( snake_case : int , snake_case : int ) -> float: """simple docstring""" if b < 0: return 1 / actual_power(snake_case , snake_case ) return actual_power(snake_case , snake_case ) if __name__ == "__main__": print(power(-2, -3))
367
'''simple docstring''' import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCamelCase : """simple docstring""" def __init__( self : List[str] , UpperCAmelCase_ : Tuple): """simple docstring""" if isinstance(UpperCAmelCase_ , UpperCAmelCase_): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden a : Dict = deepcopy(UpperCAmelCase_) elif os.path.exists(UpperCAmelCase_): with io.open(UpperCAmelCase_ , 'r' , encoding='utf-8') as f: a : Union[str, Any] = json.load(UpperCAmelCase_) else: try: a : Union[str, Any] = baseaa.urlsafe_baadecode(UpperCAmelCase_).decode('utf-8') a : List[str] = json.loads(UpperCAmelCase_) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""") a : Optional[int] = config self.set_stage_and_offload() def SCREAMING_SNAKE_CASE_ ( self : List[Any]): """simple docstring""" a : str = self.get_value('zero_optimization.stage' , -1) # offload a : Any = False if self.is_zeroa() or self.is_zeroa(): a : Tuple = set(['cpu', 'nvme']) a : int = set( [ self.get_value('zero_optimization.offload_optimizer.device'), self.get_value('zero_optimization.offload_param.device'), ]) if len(offload_devices & offload_devices_valid) > 0: a : List[str] = True def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , UpperCAmelCase_ : Dict): """simple docstring""" a : List[str] = self.config # find the config node of interest if it exists a : int = ds_key_long.split('.') a : Union[str, Any] = nodes.pop() for node in nodes: a : Union[str, Any] = config.get(UpperCAmelCase_) if config is None: return None, ds_key return config, ds_key def SCREAMING_SNAKE_CASE_ ( self : Any , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : int=None): """simple docstring""" a , a : int = self.find_config_node(UpperCAmelCase_) if config is None: return default return config.get(UpperCAmelCase_ , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any=False): """simple docstring""" a : Any = self.config # find the config node of interest if it exists a : Optional[Any] = ds_key_long.split('.') for node in nodes: a : List[str] = config a : int = config.get(UpperCAmelCase_) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""") else: return # if found remove it if parent_config is not None: parent_config.pop(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : str , UpperCAmelCase_ : str): """simple docstring""" a : List[str] = self.get_value(UpperCAmelCase_) return False if value is None else bool(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : List[str] , UpperCAmelCase_ : Union[str, Any]): """simple docstring""" a : List[Any] = self.get_value(UpperCAmelCase_) return False if value is None else not bool(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any]): """simple docstring""" return self._stage == 2 def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any]): """simple docstring""" return self._stage == 3 def SCREAMING_SNAKE_CASE_ ( self : Dict): """simple docstring""" return self._offload class UpperCamelCase : """simple docstring""" def __init__( self : str , UpperCAmelCase_ : int): """simple docstring""" a : Union[str, Any] = engine def SCREAMING_SNAKE_CASE_ ( self : List[str] , UpperCAmelCase_ : Any , **UpperCAmelCase_ : List[Any]): """simple docstring""" self.engine.backward(UpperCAmelCase_ , **UpperCAmelCase_) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCamelCase ( a_ ): """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase_ : Any): """simple docstring""" super().__init__(UpperCAmelCase_ , device_placement=UpperCAmelCase_ , scaler=UpperCAmelCase_) a : List[str] = hasattr(self.optimizer , 'overflow') def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , UpperCAmelCase_ : Dict=None): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def SCREAMING_SNAKE_CASE_ ( self : Tuple): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def SCREAMING_SNAKE_CASE_ ( self : Dict): """simple docstring""" if self.__has_overflow__: return self.optimizer.overflow return False class UpperCamelCase ( a_ ): """simple docstring""" def __init__( self : str , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[Any]): """simple docstring""" super().__init__(UpperCAmelCase_ , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : int): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCamelCase : """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int]=0.0_01 , UpperCAmelCase_ : List[Any]=0 , **UpperCAmelCase_ : Union[str, Any]): """simple docstring""" a : int = params a : str = lr a : Tuple = weight_decay a : Dict = kwargs class UpperCamelCase : """simple docstring""" def __init__( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : Any=None , UpperCAmelCase_ : Union[str, Any]=0 , **UpperCAmelCase_ : List[Any]): """simple docstring""" a : str = optimizer a : Tuple = total_num_steps a : Optional[Any] = warmup_num_steps a : List[str] = kwargs
345
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor __a = logging.get_logger(__name__) class UpperCAmelCase_ ( _a ): """simple docstring""" def __init__( self : Tuple , *snake_case_ : Union[str, Any] , **snake_case_ : Any ): warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
35
from ..utils import DummyObject, requires_backends class UpperCamelCase__ ( metaclass=__SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ =["torch", "scipy"] def __init__( self , *_A , **_A ) -> Tuple: requires_backends(self , ['''torch''', '''scipy'''] ) @classmethod def _UpperCamelCase ( cls , *_A , **_A ) -> Any: requires_backends(cls , ['''torch''', '''scipy'''] ) @classmethod def _UpperCamelCase ( cls , *_A , **_A ) -> Tuple: requires_backends(cls , ['''torch''', '''scipy'''] )
299
0
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def lowercase( UpperCamelCase_ , UpperCamelCase_=0.9_9_9 , UpperCamelCase_="cosine" , ) -> Union[str, Any]: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(UpperCamelCase_ ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(UpperCamelCase_ ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCamelCase = [] for i in range(UpperCamelCase_ ): UpperCamelCase = i / num_diffusion_timesteps UpperCamelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(UpperCamelCase_ ) / alpha_bar_fn(UpperCamelCase_ ) , UpperCamelCase_ ) ) return torch.tensor(UpperCamelCase_ , dtype=torch.floataa ) class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , __lowerCAmelCase ): __lowerCAmelCase = [e.name for e in KarrasDiffusionSchedulers] __lowerCAmelCase = 2 @register_to_config def __init__( self : List[str] , lowerCamelCase_ : int = 1000 , lowerCamelCase_ : float = 0.0_0_0_8_5 , lowerCamelCase_ : float = 0.0_1_2 , lowerCamelCase_ : str = "linear" , lowerCamelCase_ : Optional[Union[np.ndarray, List[float]]] = None , lowerCamelCase_ : str = "epsilon" , lowerCamelCase_ : str = "linspace" , lowerCamelCase_ : int = 0 , ): """simple docstring""" if trained_betas is not None: UpperCamelCase = torch.tensor(lowerCamelCase_ , dtype=torch.floataa ) elif beta_schedule == "linear": UpperCamelCase = torch.linspace(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. UpperCamelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , lowerCamelCase_ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule UpperCamelCase = betas_for_alpha_bar(lowerCamelCase_ ) else: raise NotImplementedError(f"""{beta_schedule} does is not implemented for {self.__class__}""" ) UpperCamelCase = 1.0 - self.betas UpperCamelCase = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : List[Any] , lowerCamelCase_ : List[Any]=None ): """simple docstring""" if schedule_timesteps is None: UpperCamelCase = self.timesteps UpperCamelCase = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: UpperCamelCase = 1 if len(lowerCamelCase_ ) > 1 else 0 else: UpperCamelCase = timestep.cpu().item() if torch.is_tensor(lowerCamelCase_ ) else timestep UpperCamelCase = self._index_counter[timestep_int] return indices[pos].item() @property def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def lowerCamelCase_ ( self : str , lowerCamelCase_ : torch.FloatTensor , lowerCamelCase_ : Union[float, torch.FloatTensor] , ): """simple docstring""" UpperCamelCase = self.index_for_timestep(lowerCamelCase_ ) if self.state_in_first_order: UpperCamelCase = self.sigmas[step_index] else: UpperCamelCase = self.sigmas_interpol[step_index] UpperCamelCase = sample / ((sigma**2 + 1) ** 0.5) return sample def lowerCamelCase_ ( self : Any , lowerCamelCase_ : int , lowerCamelCase_ : Union[str, torch.device] = None , lowerCamelCase_ : Optional[int] = None , ): """simple docstring""" UpperCamelCase = num_inference_steps UpperCamelCase = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": UpperCamelCase = np.linspace(0 , num_train_timesteps - 1 , lowerCamelCase_ , dtype=lowerCamelCase_ )[::-1].copy() elif self.config.timestep_spacing == "leading": UpperCamelCase = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCamelCase = (np.arange(0 , lowerCamelCase_ ) * step_ratio).round()[::-1].copy().astype(lowerCamelCase_ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": UpperCamelCase = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCamelCase = (np.arange(lowerCamelCase_ , 0 , -step_ratio )).round().copy().astype(lowerCamelCase_ ) timesteps -= 1 else: raise ValueError( f"""{self.config.timestep_spacing} is not supported. Please make sure to choose one of 'linspace', 'leading' or 'trailing'.""" ) UpperCamelCase = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) UpperCamelCase = torch.from_numpy(np.log(lowerCamelCase_ ) ).to(lowerCamelCase_ ) UpperCamelCase = np.interp(lowerCamelCase_ , np.arange(0 , len(lowerCamelCase_ ) ) , lowerCamelCase_ ) UpperCamelCase = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) UpperCamelCase = torch.from_numpy(lowerCamelCase_ ).to(device=lowerCamelCase_ ) # interpolate sigmas UpperCamelCase = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() UpperCamelCase = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) UpperCamelCase = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(lowerCamelCase_ ).startswith("""mps""" ): # mps does not support float64 UpperCamelCase = torch.from_numpy(lowerCamelCase_ ).to(lowerCamelCase_ , dtype=torch.floataa ) else: UpperCamelCase = torch.from_numpy(lowerCamelCase_ ).to(lowerCamelCase_ ) # interpolate timesteps UpperCamelCase = self.sigma_to_t(lowerCamelCase_ ).to(lowerCamelCase_ , dtype=timesteps.dtype ) UpperCamelCase = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() UpperCamelCase = torch.cat([timesteps[:1], interleaved_timesteps] ) UpperCamelCase = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter UpperCamelCase = defaultdict(lowerCamelCase_ ) def lowerCamelCase_ ( self : str , lowerCamelCase_ : List[Any] ): """simple docstring""" UpperCamelCase = sigma.log() # get distribution UpperCamelCase = log_sigma - self.log_sigmas[:, None] # get sigmas range UpperCamelCase = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) UpperCamelCase = low_idx + 1 UpperCamelCase = self.log_sigmas[low_idx] UpperCamelCase = self.log_sigmas[high_idx] # interpolate sigmas UpperCamelCase = (low - log_sigma) / (low - high) UpperCamelCase = w.clamp(0 , 1 ) # transform interpolation to time range UpperCamelCase = (1 - w) * low_idx + w * high_idx UpperCamelCase = t.view(sigma.shape ) return t @property def lowerCamelCase_ ( self : int ): """simple docstring""" return self.sample is None def lowerCamelCase_ ( self : str , lowerCamelCase_ : Union[torch.FloatTensor, np.ndarray] , lowerCamelCase_ : Union[float, torch.FloatTensor] , lowerCamelCase_ : Union[torch.FloatTensor, np.ndarray] , lowerCamelCase_ : bool = True , ): """simple docstring""" UpperCamelCase = self.index_for_timestep(lowerCamelCase_ ) # advance index counter by 1 UpperCamelCase = timestep.cpu().item() if torch.is_tensor(lowerCamelCase_ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: UpperCamelCase = self.sigmas[step_index] UpperCamelCase = self.sigmas_interpol[step_index + 1] UpperCamelCase = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method UpperCamelCase = self.sigmas[step_index - 1] UpperCamelCase = self.sigmas_interpol[step_index] UpperCamelCase = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API UpperCamelCase = 0 UpperCamelCase = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": UpperCamelCase = sigma_hat if self.state_in_first_order else sigma_interpol UpperCamelCase = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": UpperCamelCase = sigma_hat if self.state_in_first_order else sigma_interpol UpperCamelCase = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError("""prediction_type not implemented yet: sample""" ) else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`""" ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order UpperCamelCase = (sample - pred_original_sample) / sigma_hat # 3. delta timestep UpperCamelCase = sigma_interpol - sigma_hat # store for 2nd order step UpperCamelCase = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order UpperCamelCase = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep UpperCamelCase = sigma_next - sigma_hat UpperCamelCase = self.sample UpperCamelCase = None UpperCamelCase = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=lowerCamelCase_ ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : torch.FloatTensor , lowerCamelCase_ : torch.FloatTensor , lowerCamelCase_ : torch.FloatTensor , ): """simple docstring""" UpperCamelCase = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(lowerCamelCase_ ): # mps does not support float64 UpperCamelCase = self.timesteps.to(original_samples.device , dtype=torch.floataa ) UpperCamelCase = timesteps.to(original_samples.device , dtype=torch.floataa ) else: UpperCamelCase = self.timesteps.to(original_samples.device ) UpperCamelCase = timesteps.to(original_samples.device ) UpperCamelCase = [self.index_for_timestep(lowerCamelCase_ , lowerCamelCase_ ) for t in timesteps] UpperCamelCase = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): UpperCamelCase = sigma.unsqueeze(-1 ) UpperCamelCase = original_samples + noise * sigma return noisy_samples def __len__( self : Optional[int] ): """simple docstring""" return self.config.num_train_timesteps
165
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = TypeVar("""DatasetType""", Dataset, IterableDataset) def lowercase( UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = "first_exhausted" , ) -> DatasetType: '''simple docstring''' from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError("""Unable to interleave an empty list of datasets.""" ) for i, dataset in enumerate(UpperCamelCase_ ): if not isinstance(UpperCamelCase_ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ """is an empty dataset dictionary.""" ) raise ValueError( f"""Dataset at position {i} has at least one split: {list(UpperCamelCase_ )}\n""" f"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase_ ) )}']""" ) raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase_ ).__name__}.""" ) if i == 0: UpperCamelCase , UpperCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError( f"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(f"""{stopping_strategy} is not supported. Please enter a valid stopping_strategy.""" ) if dataset_type is Dataset: return _interleave_map_style_datasets( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , stopping_strategy=UpperCamelCase_ ) else: return _interleave_iterable_datasets( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , stopping_strategy=UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = 0 , ) -> DatasetType: '''simple docstring''' if not dsets: raise ValueError("""Unable to concatenate an empty list of datasets.""" ) for i, dataset in enumerate(UpperCamelCase_ ): if not isinstance(UpperCamelCase_ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ """is an empty dataset dictionary.""" ) raise ValueError( f"""Dataset at position {i} has at least one split: {list(UpperCamelCase_ )}\n""" f"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase_ ) )}']""" ) raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase_ ).__name__}.""" ) if i == 0: UpperCamelCase , UpperCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError( f"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if dataset_type is Dataset: return _concatenate_map_style_datasets(UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , axis=UpperCamelCase_ ) else: return _concatenate_iterable_datasets(UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , axis=UpperCamelCase_ )
165
1
def __UpperCamelCase ( lowerCAmelCase__ : Dict ): __a , __a : List[Any] = [], [] while len(lowerCAmelCase__ ) > 1: __a , __a : Union[str, Any] = min(lowerCAmelCase__ ), max(lowerCAmelCase__ ) start.append(lowerCAmelCase__ ) end.append(lowerCAmelCase__ ) collection.remove(lowerCAmelCase__ ) collection.remove(lowerCAmelCase__ ) end.reverse() return start + collection + end if __name__ == "__main__": lowercase__ =input('Enter numbers separated by a comma:\n').strip() lowercase__ =[int(item) for item in user_input.split(',')] print(*merge_sort(unsorted), sep=',')
216
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency lowercase__ ={ 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } lowercase__ ='ETAOINSHRDLCUMWFGYPBVKJXQZ' lowercase__ ='ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __UpperCamelCase ( lowerCAmelCase__ : str ): __a : List[Any] = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __UpperCamelCase ( lowerCAmelCase__ : tuple ): return x[0] def __UpperCamelCase ( lowerCAmelCase__ : str ): __a : Optional[Any] = get_letter_count(lowerCAmelCase__ ) __a : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(lowerCAmelCase__ ) __a : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=lowerCAmelCase__ ) __a : int = ''''''.join(freq_to_letter[freq] ) __a : Any = list(freq_to_letter_str.items() ) freq_pairs.sort(key=lowerCAmelCase__ , reverse=lowerCAmelCase__ ) __a : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(lowerCAmelCase__ ) def __UpperCamelCase ( lowerCAmelCase__ : str ): __a : int = get_frequency_order(lowerCAmelCase__ ) __a : str = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
216
1
'''simple docstring''' import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class A__ ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : Tuple ) -> str: """simple docstring""" _UpperCAmelCase : Dict = logging.get_logger() # the current default level is logging.WARNING _UpperCAmelCase : Dict = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(lowerCAmelCase__ ) def _lowerCAmelCase ( self : Optional[Any] ) -> int: """simple docstring""" _UpperCAmelCase : Any = logging.get_verbosity() _UpperCAmelCase : str = logging.get_logger("transformers.models.bart.tokenization_bart" ) _UpperCAmelCase : str = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(lowerCAmelCase__ ) as cl: logger.warning(lowerCAmelCase__ ) self.assertEqual(cl.out , msg + "\n" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(lowerCAmelCase__ ) as cl: logger.warning(lowerCAmelCase__ ) self.assertEqual(cl.out , "" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(lowerCAmelCase__ ) as cl: logger.warning(lowerCAmelCase__ ) self.assertEqual(cl.out , msg + "\n" ) # restore to the original level logging.set_verbosity(lowerCAmelCase__ ) @mockenv(TRANSFORMERS_VERBOSITY="error" ) def _lowerCAmelCase ( self : int ) -> Any: """simple docstring""" transformers.utils.logging._reset_library_root_logger() # this action activates the env var _UpperCAmelCase : Tuple = logging.get_logger("transformers.models.bart.tokenization_bart" ) _UpperCAmelCase : Union[str, Any] = os.getenv("TRANSFORMERS_VERBOSITY" , lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = logging.log_levels[env_level_str] _UpperCAmelCase : Optional[int] = logging.get_verbosity() self.assertEqual( lowerCAmelCase__ , lowerCAmelCase__ , F"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""" , ) # restore to the original level _UpperCAmelCase : Optional[Any] = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error" ) def _lowerCAmelCase ( self : int ) -> Optional[Any]: """simple docstring""" transformers.utils.logging._reset_library_root_logger() _UpperCAmelCase : Tuple = logging.logging.getLogger() with CaptureLogger(lowerCAmelCase__ ) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart" ) self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error" , cl.out ) # no need to restore as nothing was changed def _lowerCAmelCase ( self : str ) -> Optional[int]: """simple docstring""" transformers.utils.logging._reset_library_root_logger() _UpperCAmelCase : int = logging.get_logger("transformers.models.bart.tokenization_bart" ) _UpperCAmelCase : Optional[Any] = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1" ): # nothing should be logged as env var disables this method with CaptureLogger(lowerCAmelCase__ ) as cl: logger.warning_advice(lowerCAmelCase__ ) self.assertEqual(cl.out , "" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(lowerCAmelCase__ ) as cl: logger.warning_advice(lowerCAmelCase__ ) self.assertEqual(cl.out , msg + "\n" ) def __UpperCAmelCase ( ): disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
17
'''simple docstring''' import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="session" ) def __UpperCAmelCase ( ): _UpperCAmelCase : Optional[Any] = 10 _UpperCAmelCase : int = datasets.Features( { "tokens": datasets.Sequence(datasets.Value("string" ) ), "labels": datasets.Sequence(datasets.ClassLabel(names=["negative", "positive"] ) ), "answers": datasets.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), "id": datasets.Value("int64" ), } ) _UpperCAmelCase : List[str] = datasets.Dataset.from_dict( { "tokens": [["foo"] * 5] * n, "labels": [[1] * 5] * n, "answers": [{"answer_start": [97], "text": ["1976"]}] * 10, "id": list(range(a_ ) ), }, features=a_, ) return dataset @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[int], a_: Dict ): _UpperCAmelCase : Any = str(tmp_path_factory.mktemp("data" ) / "file.arrow" ) dataset.map(cache_file_name=a_ ) return filename # FILE_CONTENT + files __a = '\\n Text data.\n Second line of data.' @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Dict ): _UpperCAmelCase : Dict = tmp_path_factory.mktemp("data" ) / "file.txt" _UpperCAmelCase : Tuple = FILE_CONTENT with open(a_, "w" ) as f: f.write(a_ ) return filename @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): import bza _UpperCAmelCase : str = tmp_path_factory.mktemp("data" ) / "file.txt.bz2" _UpperCAmelCase : Optional[int] = bytes(a_, "utf-8" ) with bza.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): import gzip _UpperCAmelCase : str = str(tmp_path_factory.mktemp("data" ) / "file.txt.gz" ) _UpperCAmelCase : Any = bytes(a_, "utf-8" ) with gzip.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: str ): if datasets.config.LZ4_AVAILABLE: import lza.frame _UpperCAmelCase : Optional[int] = tmp_path_factory.mktemp("data" ) / "file.txt.lz4" _UpperCAmelCase : str = bytes(a_, "utf-8" ) with lza.frame.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: int, a_: Any ): if datasets.config.PY7ZR_AVAILABLE: import pyazr _UpperCAmelCase : Any = tmp_path_factory.mktemp("data" ) / "file.txt.7z" with pyazr.SevenZipFile(a_, "w" ) as archive: archive.write(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any, a_: List[str] ): import tarfile _UpperCAmelCase : Union[str, Any] = tmp_path_factory.mktemp("data" ) / "file.txt.tar" with tarfile.TarFile(a_, "w" ) as f: f.add(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: int ): import lzma _UpperCAmelCase : List[Any] = tmp_path_factory.mktemp("data" ) / "file.txt.xz" _UpperCAmelCase : List[str] = bytes(a_, "utf-8" ) with lzma.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Dict, a_: Tuple ): import zipfile _UpperCAmelCase : Tuple = tmp_path_factory.mktemp("data" ) / "file.txt.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[int] ): if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _UpperCAmelCase : Optional[int] = tmp_path_factory.mktemp("data" ) / "file.txt.zst" _UpperCAmelCase : int = bytes(a_, "utf-8" ) with zstd.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[int] ): _UpperCAmelCase : List[str] = tmp_path_factory.mktemp("data" ) / "file.xml" _UpperCAmelCase : Tuple = textwrap.dedent( "\\n <?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n <tmx version=\"1.4\">\n <header segtype=\"sentence\" srclang=\"ca\" />\n <body>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>" ) with open(a_, "w" ) as f: f.write(a_ ) return filename __a = [ {'col_1': '0', 'col_2': 0, 'col_3': 0.0}, {'col_1': '1', 'col_2': 1, 'col_3': 1.0}, {'col_1': '2', 'col_2': 2, 'col_3': 2.0}, {'col_1': '3', 'col_2': 3, 'col_3': 3.0}, ] __a = [ {'col_1': '4', 'col_2': 4, 'col_3': 4.0}, {'col_1': '5', 'col_2': 5, 'col_3': 5.0}, ] __a = { 'col_1': ['0', '1', '2', '3'], 'col_2': [0, 1, 2, 3], 'col_3': [0.0, 1.0, 2.0, 3.0], } __a = [ {'col_3': 0.0, 'col_1': '0', 'col_2': 0}, {'col_3': 1.0, 'col_1': '1', 'col_2': 1}, ] __a = [ {'col_1': 's0', 'col_2': 0, 'col_3': 0.0}, {'col_1': 's1', 'col_2': 1, 'col_3': 1.0}, {'col_1': 's2', 'col_2': 2, 'col_3': 2.0}, {'col_1': 's3', 'col_2': 3, 'col_3': 3.0}, ] @pytest.fixture(scope="session" ) def __UpperCAmelCase ( ): return DATA_DICT_OF_LISTS @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): _UpperCAmelCase : str = datasets.Dataset.from_dict(a_ ) _UpperCAmelCase : Optional[int] = str(tmp_path_factory.mktemp("data" ) / "dataset.arrow" ) dataset.map(cache_file_name=a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: str ): _UpperCAmelCase : int = str(tmp_path_factory.mktemp("data" ) / "dataset.sqlite" ) with contextlib.closing(sqlitea.connect(a_ ) ) as con: _UpperCAmelCase : List[Any] = con.cursor() cur.execute("CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)" ) for item in DATA: cur.execute("INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)", tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any ): _UpperCAmelCase : Dict = str(tmp_path_factory.mktemp("data" ) / "dataset.csv" ) with open(a_, "w", newline="" ) as f: _UpperCAmelCase : Dict = csv.DictWriter(a_, fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): _UpperCAmelCase : Union[str, Any] = str(tmp_path_factory.mktemp("data" ) / "dataset2.csv" ) with open(a_, "w", newline="" ) as f: _UpperCAmelCase : Optional[int] = csv.DictWriter(a_, fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: str, a_: str ): import bza _UpperCAmelCase : str = tmp_path_factory.mktemp("data" ) / "dataset.csv.bz2" with open(a_, "rb" ) as f: _UpperCAmelCase : Any = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(a_, "wb" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[int], a_: Dict, a_: Optional[int] ): _UpperCAmelCase : List[Any] = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(a_ ) ) f.write(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: List[str], a_: Union[str, Any], a_: int ): _UpperCAmelCase : int = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(csv_path.replace(".csv", ".CSV" ) ) ) f.write(a_, arcname=os.path.basename(csva_path.replace(".csv", ".CSV" ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any, a_: Union[str, Any], a_: Tuple ): _UpperCAmelCase : Any = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.csv.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Tuple ): _UpperCAmelCase : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.parquet" ) _UpperCAmelCase : Dict = pa.schema( { "col_1": pa.string(), "col_2": pa.intaa(), "col_3": pa.floataa(), } ) with open(a_, "wb" ) as f: _UpperCAmelCase : Tuple = pq.ParquetWriter(a_, schema=a_ ) _UpperCAmelCase : Tuple = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(a_ ) )] for k in DATA[0]}, schema=a_ ) writer.write_table(a_ ) writer.close() return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any ): _UpperCAmelCase : Union[str, Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) _UpperCAmelCase : str = {"data": DATA} with open(a_, "w" ) as f: json.dump(a_, a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): _UpperCAmelCase : Optional[int] = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) _UpperCAmelCase : Dict = {"data": DATA_DICT_OF_LISTS} with open(a_, "w" ) as f: json.dump(a_, a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: int ): _UpperCAmelCase : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl" ) with open(a_, "w" ) as f: for item in DATA: f.write(json.dumps(a_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Tuple ): _UpperCAmelCase : Any = str(tmp_path_factory.mktemp("data" ) / "dataset2.jsonl" ) with open(a_, "w" ) as f: for item in DATA: f.write(json.dumps(a_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any ): _UpperCAmelCase : int = str(tmp_path_factory.mktemp("data" ) / "dataset_312.jsonl" ) with open(a_, "w" ) as f: for item in DATA_312: f.write(json.dumps(a_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[Any] ): _UpperCAmelCase : Optional[int] = str(tmp_path_factory.mktemp("data" ) / "dataset-str.jsonl" ) with open(a_, "w" ) as f: for item in DATA_STR: f.write(json.dumps(a_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any], a_: Any ): import gzip _UpperCAmelCase : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.txt.gz" ) with open(a_, "rb" ) as orig_file: with gzip.open(a_, "wb" ) as zipped_file: zipped_file.writelines(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[Any], a_: Tuple ): import gzip _UpperCAmelCase : List[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl.gz" ) with open(a_, "rb" ) as orig_file: with gzip.open(a_, "wb" ) as zipped_file: zipped_file.writelines(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Dict, a_: List[Any], a_: Union[str, Any] ): _UpperCAmelCase : Tuple = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(a_ ) ) f.write(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any], a_: Optional[int], a_: Optional[Any], a_: Dict ): _UpperCAmelCase : Dict = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.join("nested", os.path.basename(a_ ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: List[Any], a_: Optional[int], a_: List[str] ): _UpperCAmelCase : Dict = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.jsonl.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: List[Any], a_: List[Any], a_: str ): _UpperCAmelCase : Optional[Any] = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.tar" with tarfile.TarFile(a_, "w" ) as f: f.add(a_, arcname=os.path.basename(a_ ) ) f.add(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: List[str], a_: List[Any], a_: Tuple, a_: Dict ): _UpperCAmelCase : List[Any] = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.tar" with tarfile.TarFile(a_, "w" ) as f: f.add(a_, arcname=os.path.join("nested", os.path.basename(a_ ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: List[str] ): _UpperCAmelCase : List[str] = ["0", "1", "2", "3"] _UpperCAmelCase : Tuple = str(tmp_path_factory.mktemp("data" ) / "dataset.txt" ) with open(a_, "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Union[str, Any] ): _UpperCAmelCase : Dict = ["0", "1", "2", "3"] _UpperCAmelCase : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset2.txt" ) with open(a_, "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any ): _UpperCAmelCase : int = ["0", "1", "2", "3"] _UpperCAmelCase : str = tmp_path_factory.mktemp("data" ) / "dataset.abc" with open(a_, "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[Any], a_: Any, a_: Union[str, Any] ): _UpperCAmelCase : Union[str, Any] = tmp_path_factory.mktemp("data" ) / "dataset.text.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(a_ ) ) f.write(a_, arcname=os.path.basename(a_ ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[int], a_: List[Any], a_: List[Any] ): _UpperCAmelCase : List[Any] = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.text.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) f.write(a_, arcname=os.path.join("main_dir", os.path.basename(a_ ) ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Any, a_: str, a_: Tuple ): _UpperCAmelCase : List[Any] = tmp_path_factory.mktemp("data" ) / "dataset.ext.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename("unsupported.ext" ) ) f.write(a_, arcname=os.path.basename("unsupported_2.ext" ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Optional[Any] ): _UpperCAmelCase : List[str] = "\n".join(["First", "Second\u2029with Unicode new line", "Third"] ) _UpperCAmelCase : str = str(tmp_path_factory.mktemp("data" ) / "dataset_with_unicode_new_lines.txt" ) with open(a_, "w", encoding="utf-8" ) as f: f.write(a_ ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( ): return os.path.join("tests", "features", "data", "test_image_rgb.jpg" ) @pytest.fixture(scope="session" ) def __UpperCAmelCase ( ): return os.path.join("tests", "features", "data", "test_audio_44100.wav" ) @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: int, a_: Optional[Any] ): _UpperCAmelCase : Union[str, Any] = tmp_path_factory.mktemp("data" ) / "dataset.img.zip" with zipfile.ZipFile(a_, "w" ) as f: f.write(a_, arcname=os.path.basename(a_ ) ) f.write(a_, arcname=os.path.basename(a_ ).replace(".jpg", "2.jpg" ) ) return path @pytest.fixture(scope="session" ) def __UpperCAmelCase ( a_: Tuple ): _UpperCAmelCase : Optional[Any] = tmp_path_factory.mktemp("data_dir" ) (data_dir / "subdir").mkdir() with open(data_dir / "subdir" / "train.txt", "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / "subdir" / "test.txt", "w" ) as f: f.write("bar\n" * 10 ) # hidden file with open(data_dir / "subdir" / ".test.txt", "w" ) as f: f.write("bar\n" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / ".subdir" / "train.txt", "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / ".subdir" / "test.txt", "w" ) as f: f.write("bar\n" * 10 ) return data_dir
17
1
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( UpperCamelCase__ ): def __init__( self , *_a , _a=None , _a=None , **_a ) -> Dict: super().__init__(*_a , **_a ) _A : Optional[int] = eval_examples _A : Tuple = post_process_function def a__ ( self , _a = None , _a=None , _a = None , _a = "eval" , **_a , ) -> Dict[str, float]: _A : Any = gen_kwargs.copy() _A : Tuple = ( gen_kwargs["""max_length"""] if gen_kwargs.get("""max_length""" ) is not None else self.args.generation_max_length ) _A : Tuple = ( gen_kwargs["""num_beams"""] if gen_kwargs.get("""num_beams""" ) is not None else self.args.generation_num_beams ) _A : List[str] = gen_kwargs _A : Dict = self.eval_dataset if eval_dataset is None else eval_dataset _A : Optional[int] = self.get_eval_dataloader(_a ) _A : Any = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _A : List[str] = self.compute_metrics _A : Tuple = None _A : str = time.time() _A : Union[str, Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _A : Optional[int] = eval_loop( _a , description="""Evaluation""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_a , metric_key_prefix=_a , ) finally: _A : List[str] = compute_metrics _A : Tuple = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( _a , _a , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default _A : Tuple = self.post_process_function(_a , _a , _a ) _A : Optional[int] = self.compute_metrics(_a ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): _A : int = metrics.pop(_a ) metrics.update(output.metrics ) else: _A : Union[str, Any] = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_a ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) _A : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , _a ) return metrics def a__ ( self , _a , _a , _a=None , _a = "test" , **_a ) -> Tuple: _A : List[str] = gen_kwargs.copy() _A : Optional[Any] = self.get_test_dataloader(_a ) # Temporarily disable metric computation, we will do it in the loop here. _A : List[Any] = self.compute_metrics _A : int = None _A : Union[str, Any] = time.time() _A : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _A : Dict = eval_loop( _a , description="""Prediction""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_a , metric_key_prefix=_a , ) finally: _A : int = compute_metrics _A : Union[str, Any] = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( _a , _a , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output _A : Union[str, Any] = self.post_process_function(_a , _a , _a , """predict""" ) _A : Tuple = self.compute_metrics(_a ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): _A : Optional[Any] = metrics.pop(_a ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_a )
26
def lowerCAmelCase_ ( snake_case_,snake_case_ ): _enforce_args(snake_case_,snake_case_ ) if n == 0: return 0 _A : Tuple = float("""-inf""" ) for i in range(1,n + 1 ): _A : str = max( snake_case_,prices[i - 1] + naive_cut_rod_recursive(n - i,snake_case_ ) ) return max_revue def lowerCAmelCase_ ( snake_case_,snake_case_ ): _enforce_args(snake_case_,snake_case_ ) _A : Dict = [float("""-inf""" ) for _ in range(n + 1 )] return _top_down_cut_rod_recursive(snake_case_,snake_case_,snake_case_ ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: _A : List[str] = float("""-inf""" ) for i in range(1,n + 1 ): _A : Optional[Any] = max( snake_case_,prices[i - 1] + _top_down_cut_rod_recursive(n - i,snake_case_,snake_case_ ),) _A : Tuple = max_revenue return max_rev[n] def lowerCAmelCase_ ( snake_case_,snake_case_ ): _enforce_args(snake_case_,snake_case_ ) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. _A : List[Any] = [float("""-inf""" ) for _ in range(n + 1 )] _A : Any = 0 for i in range(1,n + 1 ): _A : Optional[Any] = max_rev[i] for j in range(1,i + 1 ): _A : int = max(snake_case_,prices[j - 1] + max_rev[i - j] ) _A : int = max_revenue_i return max_rev[n] def lowerCAmelCase_ ( snake_case_,snake_case_ ): if n < 0: _A : Optional[Any] = f'''n must be greater than or equal to 0. Got n = {n}''' raise ValueError(snake_case_ ) if n > len(snake_case_ ): _A : Any = ( """Each integral piece of rod must have a corresponding price. """ f'''Got n = {n} but length of prices = {len(snake_case_ )}''' ) raise ValueError(snake_case_ ) def lowerCAmelCase_ ( ): _A : Tuple = [6, 10, 12, 15, 20, 23] _A : List[Any] = len(snake_case_ ) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. _A : Any = 36 _A : List[Any] = top_down_cut_rod(snake_case_,snake_case_ ) _A : List[Any] = bottom_up_cut_rod(snake_case_,snake_case_ ) _A : Dict = naive_cut_rod_recursive(snake_case_,snake_case_ ) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
26
1
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __UpperCamelCase : List[Any] = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" inspect_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = path + '''.py''' assert script_name in os.listdir(SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def _a ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" inspect_metric(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : int = path + '''.py''' assert script_name in os.listdir(SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : str ): """simple docstring""" UpperCamelCase__ : int = get_dataset_config_info(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _a ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : str ): """simple docstring""" with pytest.raises(SCREAMING_SNAKE_CASE ): get_dataset_config_info(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def _a ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" UpperCamelCase__ : List[Any] = get_dataset_config_names(SCREAMING_SNAKE_CASE ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" UpperCamelCase__ : Optional[int] = get_dataset_infos(SCREAMING_SNAKE_CASE ) assert list(infos.keys() ) == expected_configs UpperCamelCase__ : List[str] = expected_configs[0] assert expected_config in infos UpperCamelCase__ : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _a ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" UpperCamelCase__ : Optional[Any] = get_dataset_infos(SCREAMING_SNAKE_CASE ) assert expected_config in infos UpperCamelCase__ : Optional[int] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _a ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" with pytest.raises(SCREAMING_SNAKE_CASE ): get_dataset_split_names(SCREAMING_SNAKE_CASE , config_name=SCREAMING_SNAKE_CASE )
51
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __UpperCamelCase : int = logging.get_logger(__name__) class __magic_name__ ( __lowerCAmelCase): A: str = ["pixel_values"] def __init__( self : str , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : PILImageResampling = PILImageResampling.BILINEAR , lowerCamelCase__ : bool = True , lowerCamelCase__ : Union[int, float] = 1 / 255 , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : bool = True , **lowerCamelCase__ : Any , ) -> None: '''simple docstring''' super().__init__(**lowerCamelCase__ ) UpperCamelCase__ : Optional[int] = size if size is not None else {'''shortest_edge''': 224} UpperCamelCase__ : List[str] = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ ) UpperCamelCase__ : Union[str, Any] = crop_size if crop_size is not None else {'''height''': 256, '''width''': 256} UpperCamelCase__ : Dict = get_size_dict(lowerCamelCase__ , param_name='''crop_size''' ) UpperCamelCase__ : Optional[Any] = do_resize UpperCamelCase__ : List[Any] = size UpperCamelCase__ : Optional[int] = resample UpperCamelCase__ : Optional[int] = do_rescale UpperCamelCase__ : Dict = rescale_factor UpperCamelCase__ : Optional[Any] = do_center_crop UpperCamelCase__ : int = crop_size UpperCamelCase__ : List[str] = do_flip_channel_order def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : PILImageResampling = PIL.Image.BILINEAR , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : List[str] , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ : Optional[int] = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ ) if "shortest_edge" not in size: raise ValueError(F"The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}" ) UpperCamelCase__ : int = get_resize_output_image_size(lowerCamelCase__ , size=size['''shortest_edge'''] , default_to_square=lowerCamelCase__ ) return resize(lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : int , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : List[Any] , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ : Optional[int] = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}" ) return center_crop(lowerCamelCase__ , size=(size['''height'''], size['''width''']) , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Tuple , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Union[int, float] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : Tuple , ) -> List[Any]: '''simple docstring''' return rescale(lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None ) -> np.ndarray: '''simple docstring''' return flip_channel_order(lowerCamelCase__ , data_format=lowerCamelCase__ ) def UpperCAmelCase__ ( self : Dict , lowerCamelCase__ : ImageInput , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : PILImageResampling = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : float = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Optional[Union[str, TensorType]] = None , lowerCamelCase__ : ChannelDimension = ChannelDimension.FIRST , **lowerCamelCase__ : List[Any] , ) -> PIL.Image.Image: '''simple docstring''' UpperCamelCase__ : Optional[Any] = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ : List[Any] = resample if resample is not None else self.resample UpperCamelCase__ : str = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ : Any = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase__ : List[str] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) UpperCamelCase__ : List[str] = size if size is not None else self.size UpperCamelCase__ : int = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ ) UpperCamelCase__ : Tuple = crop_size if crop_size is not None else self.crop_size UpperCamelCase__ : int = get_size_dict(lowerCamelCase__ , param_name='''crop_size''' ) UpperCamelCase__ : int = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) # All transformations expect numpy arrays. UpperCamelCase__ : Union[str, Any] = [to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: UpperCamelCase__ : Tuple = [self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_center_crop: UpperCamelCase__ : Optional[Any] = [self.center_crop(image=lowerCamelCase__ , size=lowerCamelCase__ ) for image in images] if do_rescale: UpperCamelCase__ : List[Any] = [self.rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: UpperCamelCase__ : List[Any] = [self.flip_channel_order(image=lowerCamelCase__ ) for image in images] UpperCamelCase__ : Union[str, Any] = [to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] UpperCamelCase__ : int = {'''pixel_values''': images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ ) def UpperCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[Tuple] = None ) -> Tuple: '''simple docstring''' UpperCamelCase__ : Tuple = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(lowerCamelCase__ ): UpperCamelCase__ : Tuple = target_sizes.numpy() UpperCamelCase__ : Any = [] for idx in range(len(lowerCamelCase__ ) ): UpperCamelCase__ : Optional[Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=lowerCamelCase__ ) UpperCamelCase__ : Optional[Any] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(lowerCamelCase__ ) else: UpperCamelCase__ : Dict = logits.argmax(dim=1 ) UpperCamelCase__ : Dict = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
51
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : List[Any] = logging.get_logger(__name__) _a : Any = { 'microsoft/swinv2-tiny-patch4-window8-256': ( 'https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json' ), } class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : int = "swinv2" _UpperCamelCase : List[str] = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , a__=224 , a__=4 , a__=3 , a__=96 , a__=[2, 2, 6, 2] , a__=[3, 6, 12, 24] , a__=7 , a__=4.0 , a__=True , a__=0.0 , a__=0.0 , a__=0.1 , a__="gelu" , a__=False , a__=0.0_2 , a__=1e-5 , a__=32 , **a__ , ): super().__init__(**a__ ) _lowerCAmelCase : int = image_size _lowerCAmelCase : Optional[Any] = patch_size _lowerCAmelCase : Any = num_channels _lowerCAmelCase : List[Any] = embed_dim _lowerCAmelCase : Optional[int] = depths _lowerCAmelCase : List[Any] = len(a__ ) _lowerCAmelCase : Any = num_heads _lowerCAmelCase : Tuple = window_size _lowerCAmelCase : Tuple = mlp_ratio _lowerCAmelCase : Any = qkv_bias _lowerCAmelCase : Optional[int] = hidden_dropout_prob _lowerCAmelCase : Tuple = attention_probs_dropout_prob _lowerCAmelCase : str = drop_path_rate _lowerCAmelCase : List[str] = hidden_act _lowerCAmelCase : List[str] = use_absolute_embeddings _lowerCAmelCase : Optional[int] = layer_norm_eps _lowerCAmelCase : List[Any] = initializer_range _lowerCAmelCase : Any = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowerCAmelCase : Tuple = int(embed_dim * 2 ** (len(a__ ) - 1) ) _lowerCAmelCase : Tuple = (0, 0, 0, 0)
44
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCAmelCase = '''\ Text data. Second line of data.''' UpperCAmelCase = '''file''' @pytest.fixture(scope='session' ) def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / (FILE_PATH + '.zstd') __lowercase =bytes(lowercase__, 'utf-8' ) with zstd.open(lowercase__, 'wb' ) as f: f.write(lowercase__ ) return path @pytest.fixture def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir, lowercase__ ), 'w' ) as f: f.write(lowercase__ ) return FILE_PATH @pytest.mark.parametrize('compression_format', ['gzip', 'xz', 'zstd'] ) def __UpperCamelCase ( lowercase__ : Any, lowercase__ : List[str], lowercase__ : Optional[int], lowercase__ : str, lowercase__ : int, lowercase__ : Dict ): '''simple docstring''' __lowercase ={'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_path} __lowercase =input_paths[compression_format] __lowercase =tmp_path / 'cache' __lowercase =DownloadConfig(cache_dir=lowercase__, extract_compressed_file=lowercase__ ) __lowercase =cached_path(lowercase__, download_config=lowercase__ ) with open(lowercase__ ) as f: __lowercase =f.read() with open(lowercase__ ) as f: __lowercase =f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('default_extracted', [True, False] ) @pytest.mark.parametrize('default_cache_dir', [True, False] ) def __UpperCamelCase ( lowercase__ : Union[str, Any], lowercase__ : Tuple, lowercase__ : int, lowercase__ : int, lowercase__ : Optional[int] ): '''simple docstring''' __lowercase ='custom_cache' __lowercase ='custom_extracted_dir' __lowercase =tmp_path / 'custom_extracted_path' if default_extracted: __lowercase =('downloads' if default_cache_dir else custom_cache_dir, 'extracted') else: monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_DIR', lowercase__ ) monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH', str(lowercase__ ) ) __lowercase =custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __lowercase =xz_file __lowercase =( DownloadConfig(extract_compressed_file=lowercase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir, extract_compressed_file=lowercase__ ) ) __lowercase =cached_path(lowercase__, download_config=lowercase__ ) assert Path(lowercase__ ).parent.parts[-2:] == expected def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' __lowercase =str(Path(lowercase__ ).resolve() ) assert cached_path(lowercase__ ) == text_file # relative path __lowercase =str(Path(lowercase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowercase__ ) == text_file def __UpperCamelCase ( lowercase__ : Optional[Any] ): '''simple docstring''' __lowercase =str(tmp_path.resolve() / '__missing_file__.txt' ) with pytest.raises(lowercase__ ): cached_path(lowercase__ ) # relative path __lowercase ='./__missing_file__.txt' with pytest.raises(lowercase__ ): cached_path(lowercase__ ) def __UpperCamelCase ( lowercase__ : List[str] ): '''simple docstring''' __lowercase =get_from_cache(F'''tmp://{tmpfs_file}''' ) with open(lowercase__ ) as f: __lowercase =f.read() assert output_file_content == FILE_CONTENT @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( ): '''simple docstring''' with pytest.raises(lowercase__ ): cached_path('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): http_get('https://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): http_head('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): ftp_get('ftp://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): ftp_head('ftp://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): fsspec_get('s3://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): fsspec_head('s3://huggingface.co' )
141
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase__ : Union[str, Any] =logging.getLogger(__name__) def _lowercase ( _UpperCAmelCase , _UpperCAmelCase ) -> int: return (preds == labels).mean() @dataclass class __A : __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class __A : __A = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys() )} ) __A = field(metadata={"""help""": """Should contain the data files for the task."""} ) __A = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def _lowercase ( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCamelCase =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCamelCase , lowerCamelCase , lowerCamelCase =parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , _UpperCAmelCase ) # Set seed set_seed(training_args.seed ) try: lowerCamelCase =processors[data_args.task_name]() lowerCamelCase =processor.get_labels() lowerCamelCase =len(_UpperCAmelCase ) except KeyError: raise ValueError("""Task not found: %s""" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase =AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=_UpperCAmelCase , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) lowerCamelCase =AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowerCamelCase =AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_UpperCAmelCase , cache_dir=model_args.cache_dir , ) # Get datasets lowerCamelCase =( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) lowerCamelCase =( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(_UpperCAmelCase ) -> Dict: lowerCamelCase =np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(_UpperCAmelCase , p.label_ids )} # Data collator lowerCamelCase =DataCollatorWithPadding(_UpperCAmelCase , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowerCamelCase =Trainer( model=_UpperCAmelCase , args=_UpperCAmelCase , train_dataset=_UpperCAmelCase , eval_dataset=_UpperCAmelCase , compute_metrics=_UpperCAmelCase , data_collator=_UpperCAmelCase , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowerCamelCase ={} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) lowerCamelCase =trainer.evaluate() lowerCamelCase =os.path.join(training_args.output_dir , """eval_results.txt""" ) if trainer.is_world_master(): with open(_UpperCAmelCase , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" , _UpperCAmelCase , _UpperCAmelCase ) writer.write("""%s = %s\n""" % (key, value) ) results.update(_UpperCAmelCase ) return results def _lowercase ( _UpperCAmelCase ) -> Union[str, Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging UpperCAmelCase__ : List[Any] =logging.get_logger(__name__) UpperCAmelCase__ : Dict ={'''vocab_file''': '''spiece.model'''} UpperCAmelCase__ : Dict ={ '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } UpperCAmelCase__ : List[str] ={ '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) UpperCAmelCase__ : Any =0 UpperCAmelCase__ : List[Any] =1 UpperCAmelCase__ : Union[str, Any] =2 UpperCAmelCase__ : Tuple =3 UpperCAmelCase__ : int =4 class __A ( a ): __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = """left""" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<sep>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<cls>" , UpperCAmelCase_="<mask>" , UpperCAmelCase_=["<eop>", "<eod>"] , UpperCAmelCase_ = None , **UpperCAmelCase_ , ): # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase =AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token lowerCamelCase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase_ , remove_space=UpperCAmelCase_ , keep_accents=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase_ , ) lowerCamelCase =3 lowerCamelCase =do_lower_case lowerCamelCase =remove_space lowerCamelCase =keep_accents lowerCamelCase =vocab_file lowerCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase_ ) @property def _snake_case ( self ): return len(self.sp_model ) def _snake_case ( self ): lowerCamelCase ={self.convert_ids_to_tokens(UpperCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowerCamelCase =self.__dict__.copy() lowerCamelCase =None return state def __setstate__( self , UpperCAmelCase_ ): lowerCamelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase ={} lowerCamelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _snake_case ( self , UpperCAmelCase_ ): if self.remove_space: lowerCamelCase =""" """.join(inputs.strip().split() ) else: lowerCamelCase =inputs lowerCamelCase =outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: lowerCamelCase =unicodedata.normalize("""NFKD""" , UpperCAmelCase_ ) lowerCamelCase ="""""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase_ )] ) if self.do_lower_case: lowerCamelCase =outputs.lower() return outputs def _snake_case ( self , UpperCAmelCase_ ): lowerCamelCase =self.preprocess_text(UpperCAmelCase_ ) lowerCamelCase =self.sp_model.encode(UpperCAmelCase_ , out_type=UpperCAmelCase_ ) lowerCamelCase =[] for piece in pieces: if len(UpperCAmelCase_ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowerCamelCase =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase_ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase =cur_pieces[1:] else: lowerCamelCase =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase_ ) else: new_pieces.append(UpperCAmelCase_ ) return new_pieces def _snake_case ( self , UpperCAmelCase_ ): return self.sp_model.PieceToId(UpperCAmelCase_ ) def _snake_case ( self , UpperCAmelCase_ ): return self.sp_model.IdToPiece(UpperCAmelCase_ ) def _snake_case ( self , UpperCAmelCase_ ): lowerCamelCase ="""""".join(UpperCAmelCase_ ).replace(UpperCAmelCase_ , """ """ ).strip() return out_string def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , **UpperCAmelCase_ , ): lowerCamelCase =kwargs.pop("""use_source_tokenizer""" , UpperCAmelCase_ ) lowerCamelCase =self.convert_ids_to_tokens(UpperCAmelCase_ , skip_special_tokens=UpperCAmelCase_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCamelCase =[] lowerCamelCase =[] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase_ ) ) lowerCamelCase =[] sub_texts.append(UpperCAmelCase_ ) else: current_sub_text.append(UpperCAmelCase_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase_ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowerCamelCase ="""""".join(UpperCAmelCase_ ) lowerCamelCase =( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCamelCase =self.clean_up_tokenization(UpperCAmelCase_ ) return clean_text else: return text def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCamelCase =[self.sep_token_id] lowerCamelCase =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase_ )) + [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] return ([0] * len(UpperCAmelCase_ )) + [1, 1] def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCamelCase =[self.sep_token_id] lowerCamelCase =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase =os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase_ , """wb""" ) as fi: lowerCamelCase =self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase_ ) return (out_vocab_file,)
262
1
import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) _A = "bert-base-cased" _A = "fp16" _A = "bf16" _A = [FPaa, BFaa] @require_fsdp @require_cuda class _lowerCAmelCase ( __a ): def __a ( self ) -> List[str]: super().setUp() lowerCAmelCase_ = dict( ACCELERATE_USE_FSDP="true" , MASTER_ADDR="localhost" , MASTER_PORT="10999" , RANK="0" , LOCAL_RANK="0" , WORLD_SIZE="1" , ) def __a ( self ) -> str: from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(_UpperCamelCase ): lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = f"""{i + 1}""" lowerCAmelCase_ = strategy with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1 ) ) def __a ( self ) -> Any: from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(_UpperCamelCase ): lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = prefetch_policy with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch ) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1 ) ) def __a ( self ) -> List[Any]: from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(_UpperCamelCase ): lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = state_dict_type with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1 ) ) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu ) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only ) def __a ( self ) -> str: lowerCAmelCase_ = AutoModel.from_pretrained(_UpperCamelCase ) for policy in FSDP_AUTO_WRAP_POLICY: lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = policy if policy == "TRANSFORMER_BASED_WRAP": lowerCAmelCase_ = "BertLayer" elif policy == "SIZE_BASED_WRAP": lowerCAmelCase_ = "2000" with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_UpperCamelCase ) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy ) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy ) lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = "TRANSFORMER_BASED_WRAP" lowerCAmelCase_ = "T5Layer" with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() with self.assertRaises(_UpperCamelCase ) as cm: fsdp_plugin.set_auto_wrap_policy(_UpperCamelCase ) self.assertTrue("Could not find the transformer layer class to wrap in the model." in str(cm.exception ) ) lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = "SIZE_BASED_WRAP" lowerCAmelCase_ = "0" with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_UpperCamelCase ) self.assertIsNone(fsdp_plugin.auto_wrap_policy ) def __a ( self ) -> Tuple: from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = mp_dtype with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = Accelerator() if mp_dtype == "fp16": lowerCAmelCase_ = torch.floataa elif mp_dtype == "bf16": lowerCAmelCase_ = torch.bfloataa lowerCAmelCase_ = MixedPrecision(param_dtype=_UpperCamelCase , reduce_dtype=_UpperCamelCase , buffer_dtype=_UpperCamelCase ) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , _UpperCamelCase ) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , _UpperCamelCase ) ) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler ) AcceleratorState._reset_state(_UpperCamelCase ) def __a ( self ) -> List[Any]: from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: lowerCAmelCase_ = self.dist_env.copy() lowerCAmelCase_ = str(_UpperCamelCase ).lower() with mockenv_context(**_UpperCamelCase ): lowerCAmelCase_ = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=_UpperCamelCase ) ) @require_fsdp @require_multi_gpu @slow class _lowerCAmelCase ( __a ): def __a ( self ) -> Optional[int]: super().setUp() lowerCAmelCase_ = 0.82 lowerCAmelCase_ = [ "fsdp_shard_grad_op_transformer_based_wrap", "fsdp_full_shard_transformer_based_wrap", ] lowerCAmelCase_ = { "multi_gpu_fp16": 3_200, "fsdp_shard_grad_op_transformer_based_wrap_fp16": 2_000, "fsdp_full_shard_transformer_based_wrap_fp16": 1_900, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } lowerCAmelCase_ = 160 lowerCAmelCase_ = 160 lowerCAmelCase_ = inspect.getfile(accelerate.test_utils ) lowerCAmelCase_ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "external_deps"] ) def __a ( self ) -> Tuple: lowerCAmelCase_ = os.path.join(self.test_scripts_folder , "test_performance.py" ) lowerCAmelCase_ = ["accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", "--use_fsdp"] for config in self.performance_configs: lowerCAmelCase_ = cmd.copy() for i, strategy in enumerate(_UpperCamelCase ): if strategy.lower() in config: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) break if "fp32" in config: cmd_config.append("--mixed_precision=no" ) else: cmd_config.append("--mixed_precision=fp16" ) if "cpu_offload" in config: cmd_config.append("--fsdp_offload_params=True" ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("--fsdp_transformer_layer_cls_to_wrap=BertLayer" ) elif policy == "SIZE_BASED_WRAP": cmd_config.append("--fsdp_min_num_params=2000" ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--performance_lower_bound={self.performance_lower_bound}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_UpperCamelCase , env=os.environ.copy() ) def __a ( self ) -> int: lowerCAmelCase_ = os.path.join(self.test_scripts_folder , "test_checkpointing.py" ) lowerCAmelCase_ = [ "accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", "--use_fsdp", "--mixed_precision=fp16", "--fsdp_transformer_layer_cls_to_wrap=BertLayer", ] for i, strategy in enumerate(_UpperCamelCase ): lowerCAmelCase_ = cmd.copy() cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) if strategy != "FULL_SHARD": continue lowerCAmelCase_ = len(_UpperCamelCase ) for state_dict_type in FSDP_STATE_DICT_TYPE: lowerCAmelCase_ = cmd_config[:state_dict_config_index] cmd_config.append(f"""--fsdp_state_dict_type={state_dict_type}""" ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", "--partial_train_epoch=1", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_UpperCamelCase , env=os.environ.copy() ) lowerCAmelCase_ = cmd_config[:-1] lowerCAmelCase_ = os.path.join(self.tmpdir , "epoch_0" ) cmd_config.extend( [ f"""--resume_from_checkpoint={resume_from_checkpoint}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_UpperCamelCase , env=os.environ.copy() ) def __a ( self ) -> Optional[int]: lowerCAmelCase_ = os.path.join(self.test_scripts_folder , "test_peak_memory_usage.py" ) lowerCAmelCase_ = [ "accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): lowerCAmelCase_ = cmd.copy() if "fp16" in spec: cmd_config.extend(["--mixed_precision=fp16"] ) else: cmd_config.extend(["--mixed_precision=no"] ) if "multi_gpu" in spec: continue else: cmd_config.extend(["--use_fsdp"] ) for i, strategy in enumerate(_UpperCamelCase ): if strategy.lower() in spec: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) break if "cpu_offload" in spec: cmd_config.append("--fsdp_offload_params=True" ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("--fsdp_transformer_layer_cls_to_wrap=BertLayer" ) elif policy == "SIZE_BASED_WRAP": cmd_config.append("--fsdp_min_num_params=2000" ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--peak_memory_upper_bound={peak_mem_upper_bound}""", f"""--n_train={self.n_train}""", f"""--n_val={self.n_val}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_UpperCamelCase , env=os.environ.copy() )
231
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class _lowerCAmelCase ( __a ): _lowercase ='''megatron-bert''' def __init__( self , _UpperCamelCase=29_056 , _UpperCamelCase=1_024 , _UpperCamelCase=24 , _UpperCamelCase=16 , _UpperCamelCase=4_096 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=512 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=1e-1_2 , _UpperCamelCase=0 , _UpperCamelCase="absolute" , _UpperCamelCase=True , **_UpperCamelCase , ) -> int: super().__init__(pad_token_id=_UpperCamelCase , **_UpperCamelCase ) lowerCAmelCase_ = vocab_size lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = hidden_act lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = type_vocab_size lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = position_embedding_type lowerCAmelCase_ = use_cache
231
1
import math def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase): return math.pow(_UpperCAmelCase , 2) - a def lowerCamelCase__ (_UpperCAmelCase): return 2 * x def lowerCamelCase__ (_UpperCAmelCase): SCREAMING_SNAKE_CASE = 2.0 while start <= a: SCREAMING_SNAKE_CASE = math.pow(_UpperCAmelCase , 2) return start def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase = 9999 , _UpperCAmelCase = 0.00_00_00_00_00_00_01): if a < 0: raise ValueError('math domain error') SCREAMING_SNAKE_CASE = get_initial_point(_UpperCAmelCase) for _ in range(_UpperCAmelCase): SCREAMING_SNAKE_CASE = value SCREAMING_SNAKE_CASE = value - fx(_UpperCAmelCase , _UpperCAmelCase) / fx_derivative(_UpperCAmelCase) if abs(prev_value - value) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
369
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : def __init__( self , a , a=3 , a=32 , a=3 , a=10 , a=[10, 20, 30, 40] , a=[1, 1, 2, 1] , a=True , a=True , a="relu" , a=3 , a=None , ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = embeddings_size SCREAMING_SNAKE_CASE = hidden_sizes SCREAMING_SNAKE_CASE = depths SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = num_labels SCREAMING_SNAKE_CASE = scope SCREAMING_SNAKE_CASE = len(a) def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) SCREAMING_SNAKE_CASE = None if self.use_labels: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_labels) SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self) -> Union[str, Any]: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def SCREAMING_SNAKE_CASE__ ( self , a , a , a) -> Any: SCREAMING_SNAKE_CASE = TFResNetModel(config=a) SCREAMING_SNAKE_CASE = model(a) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def SCREAMING_SNAKE_CASE__ ( self , a , a , a) -> int: SCREAMING_SNAKE_CASE = self.num_labels SCREAMING_SNAKE_CASE = TFResNetForImageClassification(a) SCREAMING_SNAKE_CASE = model(a , labels=a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class _snake_case ( A__ , A__ , unittest.TestCase ): _lowercase : List[Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () _lowercase : Dict = ( {'''feature-extraction''': TFResNetModel, '''image-classification''': TFResNetForImageClassification} if is_tf_available() else {} ) _lowercase : Union[str, Any] = False _lowercase : Any = False _lowercase : List[str] = False _lowercase : str = False _lowercase : int = False def SCREAMING_SNAKE_CASE__ ( self) -> Dict: SCREAMING_SNAKE_CASE = TFResNetModelTester(self) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=a , has_text_modality=a) def SCREAMING_SNAKE_CASE__ ( self) -> Dict: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: return @unittest.skip(reason='ResNet does not use inputs_embeds') def SCREAMING_SNAKE_CASE__ ( self) -> int: pass @unittest.skip(reason='ResNet does not support input and output embeddings') def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: pass def SCREAMING_SNAKE_CASE__ ( self) -> Union[str, Any]: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(a) SCREAMING_SNAKE_CASE = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE = ['pixel_values'] self.assertListEqual(arg_names[:1] , a) def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a) def SCREAMING_SNAKE_CASE__ ( self) -> List[Any]: def check_hidden_states_output(a , a , a): SCREAMING_SNAKE_CASE = model_class(a) SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(a , a)) SCREAMING_SNAKE_CASE = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states SCREAMING_SNAKE_CASE = self.model_tester.num_stages self.assertEqual(len(a) , expected_num_stages + 1) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: SCREAMING_SNAKE_CASE = layer_type SCREAMING_SNAKE_CASE = True check_hidden_states_output(a , a , a) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE = True check_hidden_states_output(a , a , a) def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a) @slow def SCREAMING_SNAKE_CASE__ ( self) -> str: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE = TFResNetModel.from_pretrained(a) self.assertIsNotNone(a) def lowerCamelCase__ (): SCREAMING_SNAKE_CASE = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png') return image @require_tf @require_vision class _snake_case ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE__ ( self) -> Union[str, Any]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: SCREAMING_SNAKE_CASE = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=a , return_tensors='tf') # forward pass SCREAMING_SNAKE_CASE = model(**a) # verify the logits SCREAMING_SNAKE_CASE = tf.TensorShape((1, 1000)) self.assertEqual(outputs.logits.shape , a) SCREAMING_SNAKE_CASE = tf.constant([-11.10_69, -9.78_77, -8.37_77]) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , a , atol=1E-4))
327
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig snake_case_ = logging.get_logger(__name__) # General docstring snake_case_ = '''ResNetConfig''' # Base docstring snake_case_ = '''microsoft/resnet-50''' snake_case_ = [1, 2_048, 7, 7] # Image classification docstring snake_case_ = '''microsoft/resnet-50''' snake_case_ = '''tiger cat''' snake_case_ = [ '''microsoft/resnet-50''', # See all resnet models at https://huggingface.co/models?filter=resnet ] class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a , a , a = 3 , a = 1 , a = "relu"): super().__init__() lowercase__ : List[Any] = nn.Convad( a , a , kernel_size=a , stride=a , padding=kernel_size // 2 , bias=a) lowercase__ : Tuple = nn.BatchNormad(a) lowercase__ : Dict = ACTaFN[activation] if activation is not None else nn.Identity() def snake_case_ ( self , a): lowercase__ : Any = self.convolution(a) lowercase__ : Union[str, Any] = self.normalization(a) lowercase__ : Optional[Any] = self.activation(a) return hidden_state class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a): super().__init__() lowercase__ : str = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act) lowercase__ : Optional[Any] = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1) lowercase__ : str = config.num_channels def snake_case_ ( self , a): lowercase__ : List[Any] = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.') lowercase__ : Dict = self.embedder(a) lowercase__ : Optional[int] = self.pooler(a) return embedding class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a , a , a = 2): super().__init__() lowercase__ : Any = nn.Convad(a , a , kernel_size=1 , stride=a , bias=a) lowercase__ : str = nn.BatchNormad(a) def snake_case_ ( self , a): lowercase__ : Union[str, Any] = self.convolution(a) lowercase__ : Union[str, Any] = self.normalization(a) return hidden_state class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a , a , a = 1 , a = "relu"): super().__init__() lowercase__ : Any = in_channels != out_channels or stride != 1 lowercase__ : Optional[int] = ( ResNetShortCut(a , a , stride=a) if should_apply_shortcut else nn.Identity() ) lowercase__ : List[str] = nn.Sequential( ResNetConvLayer(a , a , stride=a) , ResNetConvLayer(a , a , activation=a) , ) lowercase__ : int = ACTaFN[activation] def snake_case_ ( self , a): lowercase__ : Optional[int] = hidden_state lowercase__ : Any = self.layer(a) lowercase__ : List[str] = self.shortcut(a) hidden_state += residual lowercase__ : Dict = self.activation(a) return hidden_state class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a , a , a = 1 , a = "relu" , a = 4): super().__init__() lowercase__ : Dict = in_channels != out_channels or stride != 1 lowercase__ : int = out_channels // reduction lowercase__ : Union[str, Any] = ( ResNetShortCut(a , a , stride=a) if should_apply_shortcut else nn.Identity() ) lowercase__ : Union[str, Any] = nn.Sequential( ResNetConvLayer(a , a , kernel_size=1) , ResNetConvLayer(a , a , stride=a) , ResNetConvLayer(a , a , kernel_size=1 , activation=a) , ) lowercase__ : List[str] = ACTaFN[activation] def snake_case_ ( self , a): lowercase__ : int = hidden_state lowercase__ : Tuple = self.layer(a) lowercase__ : Dict = self.shortcut(a) hidden_state += residual lowercase__ : List[Any] = self.activation(a) return hidden_state class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a , a , a , a = 2 , a = 2 , ): super().__init__() lowercase__ : int = ResNetBottleNeckLayer if config.layer_type == 'bottleneck' else ResNetBasicLayer lowercase__ : Dict = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(a , a , stride=a , activation=config.hidden_act) , *[layer(a , a , activation=config.hidden_act) for _ in range(depth - 1)] , ) def snake_case_ ( self , a): lowercase__ : Union[str, Any] = input for layer in self.layers: lowercase__ : List[str] = layer(a) return hidden_state class SCREAMING_SNAKE_CASE__ (nn.Module ): def __init__( self , a): super().__init__() lowercase__ : Optional[int] = nn.ModuleList([]) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( a , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , )) lowercase__ : str = zip(config.hidden_sizes , config.hidden_sizes[1:]) for (in_channels, out_channels), depth in zip(a , config.depths[1:]): self.stages.append(ResNetStage(a , a , a , depth=a)) def snake_case_ ( self , a , a = False , a = True): lowercase__ : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowercase__ : Optional[int] = hidden_states + (hidden_state,) lowercase__ : List[Any] = stage_module(a) if output_hidden_states: lowercase__ : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None) return BaseModelOutputWithNoAttention( last_hidden_state=a , hidden_states=a , ) class SCREAMING_SNAKE_CASE__ (__snake_case ): __lowerCamelCase : Union[str, Any] = ResNetConfig __lowerCamelCase : Tuple = """resnet""" __lowerCamelCase : Union[str, Any] = """pixel_values""" __lowerCamelCase : int = True def snake_case_ ( self , a): if isinstance(a , nn.Convad): nn.init.kaiming_normal_(module.weight , mode='fan_out' , nonlinearity='relu') elif isinstance(a , (nn.BatchNormad, nn.GroupNorm)): nn.init.constant_(module.weight , 1) nn.init.constant_(module.bias , 0) def snake_case_ ( self , a , a=False): if isinstance(a , a): lowercase__ : Optional[Any] = value snake_case_ = r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`ResNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' snake_case_ = r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare ResNet model outputting raw features without any specific head on top.""" , __snake_case , ) class SCREAMING_SNAKE_CASE__ (__snake_case ): def __init__( self , a): super().__init__(a) lowercase__ : Tuple = config lowercase__ : Dict = ResNetEmbeddings(a) lowercase__ : Any = ResNetEncoder(a) lowercase__ : Any = nn.AdaptiveAvgPoolad((1, 1)) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(a) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=a , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def snake_case_ ( self , a , a = None , a = None): lowercase__ : Dict = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase__ : Optional[Any] = self.embedder(a) lowercase__ : Any = self.encoder( a , output_hidden_states=a , return_dict=a) lowercase__ : List[Any] = encoder_outputs[0] lowercase__ : Any = self.pooler(a) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=a , pooler_output=a , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( """ ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , __snake_case , ) class SCREAMING_SNAKE_CASE__ (__snake_case ): def __init__( self , a): super().__init__(a) lowercase__ : Optional[int] = config.num_labels lowercase__ : Union[str, Any] = ResNetModel(a) # classification head lowercase__ : List[str] = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(a) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=a , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def snake_case_ ( self , a = None , a = None , a = None , a = None , ): lowercase__ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase__ : Any = self.resnet(a , output_hidden_states=a , return_dict=a) lowercase__ : Optional[int] = outputs.pooler_output if return_dict else outputs[1] lowercase__ : Optional[Any] = self.classifier(a) lowercase__ : str = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase__ : Union[str, Any] = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase__ : Optional[Any] = 'single_label_classification' else: lowercase__ : Any = 'multi_label_classification' if self.config.problem_type == "regression": lowercase__ : List[Any] = MSELoss() if self.num_labels == 1: lowercase__ : List[Any] = loss_fct(logits.squeeze() , labels.squeeze()) else: lowercase__ : Dict = loss_fct(a , a) elif self.config.problem_type == "single_label_classification": lowercase__ : Tuple = CrossEntropyLoss() lowercase__ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) elif self.config.problem_type == "multi_label_classification": lowercase__ : str = BCEWithLogitsLoss() lowercase__ : int = loss_fct(a , a) if not return_dict: lowercase__ : Union[str, Any] = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=a , logits=a , hidden_states=outputs.hidden_states) @add_start_docstrings( """ ResNet backbone, to be used with frameworks like DETR and MaskFormer. """ , __snake_case , ) class SCREAMING_SNAKE_CASE__ (__snake_case , __snake_case ): def __init__( self , a): super().__init__(a) super()._init_backbone(a) lowercase__ : int = [config.embedding_size] + config.hidden_sizes lowercase__ : Optional[int] = ResNetEmbeddings(a) lowercase__ : List[Any] = ResNetEncoder(a) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(a) @replace_return_docstrings(output_type=a , config_class=_CONFIG_FOR_DOC) def snake_case_ ( self , a , a = None , a = None): lowercase__ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase__ : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__ : Optional[Any] = self.embedder(a) lowercase__ : Optional[int] = self.encoder(a , output_hidden_states=a , return_dict=a) lowercase__ : str = outputs.hidden_states lowercase__ : Union[str, Any] = () for idx, stage in enumerate(self.stage_names): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: lowercase__ : int = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=a , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=a , )
214
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
214
1
'''simple docstring''' import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__ = logging.get_logger() @dataclass class a_ : lowercase = 42 lowercase = field(default_factory=lowerCamelCase ) lowercase = field(default_factory=lowerCamelCase ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" UpperCamelCase = len(list(m.modules() ) ) == 1 or isinstance(_SCREAMING_SNAKE_CASE , nn.Convad ) or isinstance(_SCREAMING_SNAKE_CASE , nn.BatchNormad ) if has_not_submodules: self.traced.append(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(_SCREAMING_SNAKE_CASE ) [x.remove() for x in self.handles] return self @property def A__ ( self ) -> Tuple: """simple docstring""" return list(filter(lambda _SCREAMING_SNAKE_CASE : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class a_ : lowercase = 42 lowercase = 42 lowercase = 0 lowercase = field(default_factory=lowerCamelCase ) lowercase = field(default_factory=lowerCamelCase ) def __call__( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = Tracker(self.dest )(_SCREAMING_SNAKE_CASE ).parametrized UpperCamelCase = Tracker(self.src )(_SCREAMING_SNAKE_CASE ).parametrized UpperCamelCase = list(filter(lambda _SCREAMING_SNAKE_CASE : type(_SCREAMING_SNAKE_CASE ) not in self.src_skip , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase = list(filter(lambda _SCREAMING_SNAKE_CASE : type(_SCREAMING_SNAKE_CASE ) not in self.dest_skip , _SCREAMING_SNAKE_CASE ) ) if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise Exception( F"Numbers of operations are different. Source module has {len(_SCREAMING_SNAKE_CASE )} operations while" F" destination module has {len(_SCREAMING_SNAKE_CASE )}." ) for dest_m, src_m in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F"Transfered from={src_m} to={dest_m}" ) def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = True )-> Optional[Any]: print(F"Converting {name}..." ) with torch.no_grad(): UpperCamelCase = timm.create_model(__UpperCamelCase , pretrained=__UpperCamelCase ).eval() UpperCamelCase = ResNetForImageClassification(__UpperCamelCase ).eval() UpperCamelCase = ModuleTransfer(src=__UpperCamelCase , dest=__UpperCamelCase ) UpperCamelCase = torch.randn((1, 3, 224, 224) ) module_transfer(__UpperCamelCase ) assert torch.allclose(from_model(__UpperCamelCase ) , our_model(__UpperCamelCase ).logits ), "The model logits don't match the original one." UpperCamelCase = F"resnet{'-'.join(name.split('resnet' ) )}" print(__UpperCamelCase ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message="""Add model""" , use_temp_dir=__UpperCamelCase , ) # we can use the convnext one UpperCamelCase = AutoImageProcessor.from_pretrained("""facebook/convnext-base-224-22k-1k""" ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message="""Add image processor""" , use_temp_dir=__UpperCamelCase , ) print(F"Pushed {checkpoint_name}" ) def lowercase__ ( __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = True )-> str: UpperCamelCase = """imagenet-1k-id2label.json""" UpperCamelCase = 1000 UpperCamelCase = (1, num_labels) UpperCamelCase = """huggingface/label-files""" UpperCamelCase = num_labels UpperCamelCase = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCamelCase = {int(__UpperCamelCase ): v for k, v in idalabel.items()} UpperCamelCase = idalabel UpperCamelCase = {v: k for k, v in idalabel.items()} UpperCamelCase = partial(__UpperCamelCase , num_labels=__UpperCamelCase , idalabel=__UpperCamelCase , labelaid=__UpperCamelCase ) UpperCamelCase = { """resnet18""": ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[64, 128, 256, 512] , layer_type="""basic""" ), """resnet26""": ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[256, 512, 1024, 2048] , layer_type="""bottleneck""" ), """resnet34""": ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[64, 128, 256, 512] , layer_type="""basic""" ), """resnet50""": ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type="""bottleneck""" ), """resnet101""": ImageNetPreTrainedConfig( depths=[3, 4, 23, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type="""bottleneck""" ), """resnet152""": ImageNetPreTrainedConfig( depths=[3, 8, 36, 3] , hidden_sizes=[256, 512, 1024, 2048] , layer_type="""bottleneck""" ), } if model_name: convert_weight_and_push(__UpperCamelCase , names_to_config[model_name] , __UpperCamelCase , __UpperCamelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return config, expected_shape if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported resnet* architecture,' ' currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
183
'''simple docstring''' import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a_ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=[32, 64, 128] , _SCREAMING_SNAKE_CASE=[1, 2, 1] , _SCREAMING_SNAKE_CASE=[2, 2, 4] , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=2.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=["stage1", "stage2"] , _SCREAMING_SNAKE_CASE=[1, 2] , ) -> Optional[Any]: """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = embed_dim UpperCamelCase = hidden_sizes UpperCamelCase = depths UpperCamelCase = num_heads UpperCamelCase = window_size UpperCamelCase = mlp_ratio UpperCamelCase = qkv_bias UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = drop_path_rate UpperCamelCase = hidden_act UpperCamelCase = use_absolute_embeddings UpperCamelCase = patch_norm UpperCamelCase = layer_norm_eps UpperCamelCase = initializer_range UpperCamelCase = is_training UpperCamelCase = scope UpperCamelCase = use_labels UpperCamelCase = type_sequence_label_size UpperCamelCase = encoder_stride UpperCamelCase = out_features UpperCamelCase = out_indices def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase = self.get_config() return config, pixel_values, labels def A__ ( self ) -> str: """simple docstring""" return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" UpperCamelCase = FocalNetModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) UpperCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = FocalNetBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None UpperCamelCase = None UpperCamelCase = FocalNetBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = FocalNetForMaskedImageModeling(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCamelCase = 1 UpperCamelCase = FocalNetForMaskedImageModeling(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" UpperCamelCase = self.type_sequence_label_size UpperCamelCase = FocalNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase = 1 UpperCamelCase = FocalNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase ,UpperCamelCase ,UpperCamelCase = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a_ ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) lowercase = ( {"""feature-extraction""": FocalNetModel, """image-classification""": FocalNetForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False def A__ ( self ) -> int: """simple docstring""" UpperCamelCase = FocalNetModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , embed_dim=37 , has_text_modality=_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[Any]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self ) -> Tuple: """simple docstring""" return def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""FocalNet does not use inputs_embeds""" ) def A__ ( self ) -> int: """simple docstring""" pass @unittest.skip(reason="""FocalNet does not use feedforward chunking""" ) def A__ ( self ) -> int: """simple docstring""" pass def A__ ( self ) -> str: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear ) ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCamelCase = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase = outputs.hidden_states UpperCamelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # FocalNet has a different seq_length UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) UpperCamelCase = outputs.reshaped_hidden_states self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = reshaped_hidden_states[0].shape UpperCamelCase = ( reshaped_hidden_states[0].view(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = 3 UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) @slow def A__ ( self ) -> Union[str, Any]: """simple docstring""" for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = FocalNetModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = _config_zero_init(_SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: UpperCamelCase = model_class(config=_SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @require_vision @require_torch class a_ ( unittest.TestCase ): @cached_property def A__ ( self ) -> List[str]: """simple docstring""" return AutoImageProcessor.from_pretrained("""microsoft/focalnet-tiny""" ) if is_vision_available() else None @slow def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = FocalNetForImageClassification.from_pretrained("""microsoft/focalnet-tiny""" ).to(_SCREAMING_SNAKE_CASE ) UpperCamelCase = self.default_image_processor UpperCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCamelCase = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): UpperCamelCase = model(**_SCREAMING_SNAKE_CASE ) # verify the logits UpperCamelCase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) UpperCamelCase = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class a_ ( lowerCamelCase , unittest.TestCase ): lowercase = (FocalNetBackbone,) if is_torch_available() else () lowercase = FocalNetConfig lowercase = False def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = FocalNetModelTester(self )
183
1
'''simple docstring''' def __snake_case ( UpperCAmelCase_ : int = 1000 ): lowerCamelCase_ = 2**power lowerCamelCase_ = 0 while n: lowerCamelCase_ ,lowerCamelCase_ = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
55
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType A : str = logging.get_logger(__name__) A : Union[str, Any] = { '''microsoft/layoutlmv3-base''': '''https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json''', } class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Optional[Any] = '''layoutlmv3''' def __init__( self : Tuple , __lowerCAmelCase : Optional[int]=5_02_65 , __lowerCAmelCase : Tuple=7_68 , __lowerCAmelCase : Union[str, Any]=12 , __lowerCAmelCase : Any=12 , __lowerCAmelCase : List[Any]=30_72 , __lowerCAmelCase : Optional[int]="gelu" , __lowerCAmelCase : Dict=0.1 , __lowerCAmelCase : List[Any]=0.1 , __lowerCAmelCase : Dict=5_12 , __lowerCAmelCase : Any=2 , __lowerCAmelCase : Dict=0.0_2 , __lowerCAmelCase : List[str]=1e-5 , __lowerCAmelCase : List[str]=1 , __lowerCAmelCase : int=0 , __lowerCAmelCase : Dict=2 , __lowerCAmelCase : Tuple=10_24 , __lowerCAmelCase : List[str]=1_28 , __lowerCAmelCase : Optional[int]=1_28 , __lowerCAmelCase : Any=True , __lowerCAmelCase : Optional[Any]=32 , __lowerCAmelCase : Any=1_28 , __lowerCAmelCase : str=64 , __lowerCAmelCase : Optional[int]=2_56 , __lowerCAmelCase : int=True , __lowerCAmelCase : int=True , __lowerCAmelCase : List[str]=True , __lowerCAmelCase : int=2_24 , __lowerCAmelCase : Dict=3 , __lowerCAmelCase : List[Any]=16 , __lowerCAmelCase : Dict=None , **__lowerCAmelCase : Optional[Any] , ) -> Dict: """simple docstring""" super().__init__( vocab_size=__lowerCAmelCase , hidden_size=__lowerCAmelCase , num_hidden_layers=__lowerCAmelCase , num_attention_heads=__lowerCAmelCase , intermediate_size=__lowerCAmelCase , hidden_act=__lowerCAmelCase , hidden_dropout_prob=__lowerCAmelCase , attention_probs_dropout_prob=__lowerCAmelCase , max_position_embeddings=__lowerCAmelCase , type_vocab_size=__lowerCAmelCase , initializer_range=__lowerCAmelCase , layer_norm_eps=__lowerCAmelCase , pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase , ) A__ = max_ad_position_embeddings A__ = coordinate_size A__ = shape_size A__ = has_relative_attention_bias A__ = rel_pos_bins A__ = max_rel_pos A__ = has_spatial_attention_bias A__ = rel_ad_pos_bins A__ = max_rel_ad_pos A__ = text_embed A__ = visual_embed A__ = input_size A__ = num_channels A__ = patch_size A__ = classifier_dropout class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : List[str] = version.parse('''1.12''' ) @property def a_ ( self : int ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ("""bbox""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) else: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""bbox""", {0: """batch""", 1: """sequence"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels"""}), ] ) @property def a_ ( self : Optional[int] ) -> float: """simple docstring""" return 1e-5 @property def a_ ( self : Tuple ) -> int: """simple docstring""" return 12 def a_ ( self : str , __lowerCAmelCase : "ProcessorMixin" , __lowerCAmelCase : int = -1 , __lowerCAmelCase : int = -1 , __lowerCAmelCase : bool = False , __lowerCAmelCase : Optional["TensorType"] = None , __lowerCAmelCase : int = 3 , __lowerCAmelCase : int = 40 , __lowerCAmelCase : int = 40 , ) -> Mapping[str, Any]: """simple docstring""" setattr(processor.image_processor , """apply_ocr""" , __lowerCAmelCase ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX A__ = compute_effective_axis_dimension( __lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX A__ = processor.tokenizer.num_special_tokens_to_add(__lowerCAmelCase ) A__ = compute_effective_axis_dimension( __lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowerCAmelCase ) # Generate dummy inputs according to compute batch and sequence A__ = [[""" """.join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes A__ = [[[48, 84, 73, 1_28]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) A__ = self._generate_dummy_images(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A__ = dict( processor( __lowerCAmelCase , text=__lowerCAmelCase , boxes=__lowerCAmelCase , return_tensors=__lowerCAmelCase , ) ) return inputs
274
0
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version lowercase__ :Any = logging.getLogger(__name__) require_version("pytorch_lightning>=1.0.4") lowercase__ :Optional[Any] = { "base": AutoModel, "sequence-classification": AutoModelForSequenceClassification, "question-answering": AutoModelForQuestionAnswering, "pretraining": AutoModelForPreTraining, "token-classification": AutoModelForTokenClassification, "language-modeling": AutoModelWithLMHead, "summarization": AutoModelForSeqaSeqLM, "translation": AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization lowercase__ :Tuple = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } lowercase__ :Union[str, Any] = sorted(arg_to_scheduler.keys()) lowercase__ :List[str] = "{" + ", ".join(arg_to_scheduler_choices) + "}" class lowercase ( pl.LightningModule ): def __init__( self ,A__ ,A__=None ,A__="base" ,A__=None ,A__=None ,A__=None ,**A__ ,): super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(A__) lowercase = 0 lowercase = Path(self.hparams.output_dir) lowercase = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: lowercase = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path ,**({'''num_labels''': num_labels} if num_labels is not None else {}) ,cache_dir=A__ ,**A__ ,) else: lowercase = config lowercase = ('''encoder_layerdrop''', '''decoder_layerdrop''', '''dropout''', '''attention_dropout''') for p in extra_model_params: if getattr(self.hparams ,A__ ,A__): assert hasattr(self.config ,A__), f'model config doesn\'t have a `{p}` attribute' setattr(self.config ,A__ ,getattr(self.hparams ,A__)) if tokenizer is None: lowercase = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path ,cache_dir=A__ ,) else: lowercase = tokenizer lowercase = MODEL_MODES[mode] if model is None: lowercase = self.model_type.from_pretrained( self.hparams.model_name_or_path ,from_tf=bool('''.ckpt''' in self.hparams.model_name_or_path) ,config=self.config ,cache_dir=A__ ,) else: lowercase = model def A__ ( self ,*A__ ,**A__): lowercase = self.model_type.from_pretrained(*A__ ,**A__) def A__ ( self): lowercase = arg_to_scheduler[self.hparams.lr_scheduler] lowercase = get_schedule_func( self.opt ,num_warmup_steps=self.hparams.warmup_steps ,num_training_steps=self.total_steps()) lowercase = {'''scheduler''': scheduler, '''interval''': '''step''', '''frequency''': 1} return scheduler def A__ ( self): lowercase = self.model lowercase = ['''bias''', '''LayerNorm.weight'''] lowercase = [ { '''params''': [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay) ], # check this named paramters '''weight_decay''': self.hparams.weight_decay, }, { '''params''': [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], '''weight_decay''': 0.0, }, ] if self.hparams.adafactor: lowercase = Adafactor( A__ ,lr=self.hparams.learning_rate ,scale_parameter=A__ ,relative_step=A__) else: lowercase = AdamW( A__ ,lr=self.hparams.learning_rate ,eps=self.hparams.adam_epsilon) lowercase = optimizer lowercase = self.get_lr_scheduler() return [optimizer], [scheduler] def A__ ( self ,A__ ,A__): return self.validation_step(A__ ,A__) def A__ ( self ,A__): return self.validation_end(A__) def A__ ( self): lowercase = max(1 ,self.hparams.gpus) # TODO: consider num_tpu_cores lowercase = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def A__ ( self ,A__): if stage == "test": lowercase = len(self.test_dataloader().dataset) else: lowercase = self.get_dataloader('''train''' ,self.hparams.train_batch_size ,shuffle=A__) lowercase = len(self.train_dataloader().dataset) def A__ ( self ,A__ ,A__ ,A__ = False): raise NotImplementedError('''You must implement this for your task''') def A__ ( self): return self.train_loader def A__ ( self): return self.get_dataloader('''dev''' ,self.hparams.eval_batch_size ,shuffle=A__) def A__ ( self): return self.get_dataloader('''test''' ,self.hparams.eval_batch_size ,shuffle=A__) def A__ ( self ,A__): return os.path.join( self.hparams.data_dir ,'''cached_{}_{}_{}'''.format( A__ ,list(filter(A__ ,self.hparams.model_name_or_path.split('''/'''))).pop() ,str(self.hparams.max_seq_length) ,) ,) @pl.utilities.rank_zero_only def A__ ( self ,A__): lowercase = self.output_dir.joinpath('''best_tfmr''') lowercase = self.step_count self.model.save_pretrained(A__) self.tokenizer.save_pretrained(A__) @staticmethod def A__ ( A__ ,A__): parser.add_argument( '''--model_name_or_path''' ,default=A__ ,type=A__ ,required=A__ ,help='''Path to pretrained model or model identifier from huggingface.co/models''' ,) parser.add_argument( '''--config_name''' ,default='''''' ,type=A__ ,help='''Pretrained config name or path if not the same as model_name''') parser.add_argument( '''--tokenizer_name''' ,default=A__ ,type=A__ ,help='''Pretrained tokenizer name or path if not the same as model_name''' ,) parser.add_argument( '''--cache_dir''' ,default=str(Path(A__).parent / '''test_run''' / '''cache''') ,type=A__ ,help='''Where do you want to store the pre-trained models downloaded from huggingface.co''' ,) parser.add_argument( '''--encoder_layerdrop''' ,type=A__ ,help='''Encoder layer dropout probability (Optional). Goes into model.config''' ,) parser.add_argument( '''--decoder_layerdrop''' ,type=A__ ,help='''Decoder layer dropout probability (Optional). Goes into model.config''' ,) parser.add_argument( '''--dropout''' ,type=A__ ,help='''Dropout probability (Optional). Goes into model.config''' ,) parser.add_argument( '''--attention_dropout''' ,type=A__ ,help='''Attention dropout probability (Optional). Goes into model.config''' ,) parser.add_argument('''--learning_rate''' ,default=5E-5 ,type=A__ ,help='''The initial learning rate for Adam.''') parser.add_argument( '''--lr_scheduler''' ,default='''linear''' ,choices=A__ ,metavar=A__ ,type=A__ ,help='''Learning rate scheduler''' ,) parser.add_argument('''--weight_decay''' ,default=0.0 ,type=A__ ,help='''Weight decay if we apply some.''') parser.add_argument('''--adam_epsilon''' ,default=1E-8 ,type=A__ ,help='''Epsilon for Adam optimizer.''') parser.add_argument('''--warmup_steps''' ,default=0 ,type=A__ ,help='''Linear warmup over warmup_steps.''') parser.add_argument('''--num_workers''' ,default=4 ,type=A__ ,help='''kwarg passed to DataLoader''') parser.add_argument('''--num_train_epochs''' ,dest='''max_epochs''' ,default=3 ,type=A__) parser.add_argument('''--train_batch_size''' ,default=3_2 ,type=A__) parser.add_argument('''--eval_batch_size''' ,default=3_2 ,type=A__) parser.add_argument('''--adafactor''' ,action='''store_true''') class lowercase ( pl.Callback ): def A__ ( self ,A__ ,A__): if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class lowercase ( pl.Callback ): def A__ ( self ,A__ ,A__): # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(A__) class lowercase ( pl.Callback ): def A__ ( self ,A__ ,A__): lowercase = trainer.lr_schedulers[0]['''scheduler'''] lowercase = {f'lr_group_{i}': lr for i, lr in enumerate(lr_scheduler.get_lr())} pl_module.logger.log_metrics(A__) def A__ ( self ,A__ ,A__): rank_zero_info('''***** Validation results *****''') lowercase = trainer.callback_metrics # Log results for key in sorted(A__): if key not in ["log", "progress_bar"]: rank_zero_info('''{} = {}\n'''.format(A__ ,str(metrics[key]))) def A__ ( self ,A__ ,A__): rank_zero_info('''***** Test results *****''') lowercase = trainer.callback_metrics # Log and save results to file lowercase = os.path.join(pl_module.hparams.output_dir ,'''test_results.txt''') with open(A__ ,'''w''') as writer: for key in sorted(A__): if key not in ["log", "progress_bar"]: rank_zero_info('''{} = {}\n'''.format(A__ ,str(metrics[key]))) writer.write('''{} = {}\n'''.format(A__ ,str(metrics[key]))) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' parser.add_argument( '''--output_dir''' , default=str(Path(lowerCAmelCase__ ).parent / '''test_run''' / '''model_checkpoints''' ) , type=lowerCAmelCase__ , help='''The output directory where the model predictions and checkpoints will be written.''' , ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=lowerCAmelCase__ , default='''O2''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_tpu_cores''' , dest='''tpu_cores''' , type=lowerCAmelCase__ ) parser.add_argument('''--max_grad_norm''' , dest='''gradient_clip_val''' , default=1.0 , type=lowerCAmelCase__ , help='''Max gradient norm''' ) parser.add_argument('''--do_train''' , action='''store_true''' , help='''Whether to run training.''' ) parser.add_argument('''--do_predict''' , action='''store_true''' , help='''Whether to run predictions on the test set.''' ) parser.add_argument( '''--gradient_accumulation_steps''' , dest='''accumulate_grad_batches''' , type=lowerCAmelCase__ , default=1 , help='''Number of updates steps to accumulate before performing a backward/update pass.''' , ) parser.add_argument('''--seed''' , type=lowerCAmelCase__ , default=42 , help='''random seed for initialization''' ) parser.add_argument( '''--data_dir''' , default=str(Path(lowerCAmelCase__ ).parent / '''test_run''' / '''dummy-train-data''' ) , type=lowerCAmelCase__ , help='''The input data dir. Should contain the training files for the CoNLL-2003 NER task.''' , ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=[] , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ , ): '''simple docstring''' pl.seed_everything(args.seed ) # init model lowercase = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=lowerCAmelCase__ ) # add custom checkpoints if checkpoint_callback is None: lowercase = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix='''checkpoint''' , monitor='''val_loss''' , mode='''min''' , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(lowerCAmelCase__ ) if logging_callback is None: lowercase = LoggingCallback() lowercase = {} if args.fpaa: lowercase = 16 if args.gpus > 1: lowercase = '''auto''' lowercase = '''ddp''' lowercase = args.accumulate_grad_batches lowercase = None lowercase = '''auto''' lowercase = pl.Trainer.from_argparse_args( lowerCAmelCase__ , weights_summary=lowerCAmelCase__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=lowerCAmelCase__ , val_check_interval=1 , num_sanity_val_steps=2 , **lowerCAmelCase__ , ) if args.do_train: trainer.fit(lowerCAmelCase__ ) else: print('''RAG modeling tests with new set functions successfuly executed!''' ) return trainer
360
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase__ :str = logging.get_logger(__name__) lowercase__ :Any = {"vocab_file": "sentencepiece.bpe.model"} lowercase__ :Tuple = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } lowercase__ :str = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } lowercase__ :int = "▁" class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =VOCAB_FILES_NAMES lowercase_ : List[Any] =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Tuple =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : str =['''input_ids''', '''attention_mask'''] def __init__( self ,A__ ,A__="<s>" ,A__="</s>" ,A__="</s>" ,A__="<s>" ,A__="<unk>" ,A__="<pad>" ,A__="<mask>" ,A__ = None ,**A__ ,): # Mask token behave like a normal word, i.e. include the space before it lowercase = AddedToken(A__ ,lstrip=A__ ,rstrip=A__) if isinstance(A__ ,A__) else mask_token lowercase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A__ ,eos_token=A__ ,unk_token=A__ ,sep_token=A__ ,cls_token=A__ ,pad_token=A__ ,mask_token=A__ ,sp_model_kwargs=self.sp_model_kwargs ,**A__ ,) lowercase = vocab_file lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(A__)) lowercase = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} lowercase = len(self.sp_model) - 1 lowercase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A__ ( self ,A__ ,A__ = None): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase = [self.cls_token_id] lowercase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self ,A__ ,A__ = None ,A__ = False): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A__ ,token_ids_a=A__ ,already_has_special_tokens=A__) if token_ids_a is None: return [1] + ([0] * len(A__)) + [1] return [1] + ([0] * len(A__)) + [1, 1] + ([0] * len(A__)) + [1] def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] @property def A__ ( self): return len(self.sp_model) def A__ ( self): lowercase = {self.convert_ids_to_tokens(A__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def A__ ( self ,A__): return self.sp_model.encode(A__ ,out_type=A__) def A__ ( self ,A__): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase = self.sp_model.PieceToId(A__) return spm_id if spm_id else self.unk_token_id def A__ ( self ,A__): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(A__) def A__ ( self ,A__): lowercase = [] lowercase = '''''' lowercase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A__) + token lowercase = True lowercase = [] else: current_sub_tokens.append(A__) lowercase = False out_string += self.sp_model.decode(A__) return out_string.strip() def __getstate__( self): lowercase = self.__dict__.copy() lowercase = None return state def __setstate__( self ,A__): lowercase = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs'''): lowercase = {} lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def A__ ( self ,A__ ,A__ = None): if not os.path.isdir(A__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(A__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file ,A__) elif not os.path.isfile(self.vocab_file): with open(A__ ,'''wb''') as fi: lowercase = self.sp_model.serialized_model_proto() fi.write(A__) return (out_vocab_file,)
97
0
'''simple docstring''' from __future__ import annotations from random import choice def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' return choice(lowercase__ ) def __UpperCamelCase ( lowercase__ : list[int], lowercase__ : int ): '''simple docstring''' __lowercase =random_pivot(lowercase__ ) # partition based on pivot # linear time __lowercase =[e for e in lst if e < pivot] __lowercase =[e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(lowercase__ ) == k - 1: return pivot # pivot is in elements bigger than k elif len(lowercase__ ) < k - 1: return kth_number(lowercase__, k - len(lowercase__ ) - 1 ) # pivot is in elements smaller than k else: return kth_number(lowercase__, lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
141
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class lowerCAmelCase : def __init__( self : Optional[int] , __lowercase : Dict , __lowercase : Optional[Any]=3 , __lowercase : Union[str, Any]=7 , __lowercase : Any=True , __lowercase : List[Any]=True , __lowercase : Union[str, Any]=False , __lowercase : int=True , __lowercase : List[str]=99 , __lowercase : int=32 , __lowercase : Dict=5 , __lowercase : Union[str, Any]=4 , __lowercase : List[Any]=37 , __lowercase : str="gelu" , __lowercase : int=0.1 , __lowercase : Dict=0.1 , __lowercase : Any=512 , __lowercase : List[str]=16 , __lowercase : Tuple=2 , __lowercase : Tuple=0.0_2 , __lowercase : List[str]=3 , __lowercase : Union[str, Any]=4 , __lowercase : List[Any]=None , ): """simple docstring""" __lowercase =parent __lowercase =batch_size __lowercase =seq_length __lowercase =is_training __lowercase =use_input_mask __lowercase =use_token_type_ids __lowercase =use_labels __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =intermediate_size __lowercase =hidden_act __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =type_vocab_size __lowercase =type_sequence_label_size __lowercase =initializer_range __lowercase =num_labels __lowercase =num_choices __lowercase =scope def snake_case ( self : List[Any] ): """simple docstring""" __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase =None if self.use_input_mask: __lowercase =random_attention_mask([self.batch_size, self.seq_length] ) __lowercase =None __lowercase =None __lowercase =None __lowercase =None if self.use_labels: __lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase =ids_tensor([self.batch_size] , self.num_choices ) __lowercase =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self : Tuple ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=__lowercase , ) def snake_case ( self : str , __lowercase : Optional[Any] , __lowercase : int , __lowercase : Any , __lowercase : Tuple , __lowercase : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : Optional[Any] ): """simple docstring""" __lowercase =FalconModel(config=__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase ) __lowercase =model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Optional[Any] , __lowercase : Any , __lowercase : Union[str, Any] , __lowercase : List[Any] , __lowercase : str , __lowercase : Tuple , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[Any] , __lowercase : List[str] , ): """simple docstring""" __lowercase =True __lowercase =FalconModel(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , ) __lowercase =model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , ) __lowercase =model(__lowercase , attention_mask=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Union[str, Any] , __lowercase : int , __lowercase : List[Any] , __lowercase : List[str] , __lowercase : str , __lowercase : List[Any] , __lowercase : Tuple , __lowercase : Union[str, Any] , __lowercase : List[str] , __lowercase : Optional[int] , ): """simple docstring""" __lowercase =FalconForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self : str , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : Dict , __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : List[Any] , __lowercase : Dict , __lowercase : Tuple , ): """simple docstring""" __lowercase =True __lowercase =True __lowercase =FalconForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() # first forward pass __lowercase =model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , use_cache=__lowercase , ) __lowercase =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __lowercase =ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowercase =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __lowercase =torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase =torch.cat([input_mask, next_mask] , dim=-1 ) __lowercase =model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , output_hidden_states=__lowercase , )['hidden_states'][0] __lowercase =model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , past_key_values=__lowercase , output_hidden_states=__lowercase , )['hidden_states'][0] # select random slice __lowercase =ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase =output_from_no_past[:, -3:, random_slice_idx].detach() __lowercase =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3 ) ) def snake_case ( self : Optional[Any] ): """simple docstring""" __lowercase =self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) =config_and_inputs __lowercase ={'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowerCAmelCase ( A , A , A , unittest.TestCase ): lowerCAmelCase_ = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) lowerCAmelCase_ = (FalconForCausalLM,) if is_torch_available() else () lowerCAmelCase_ = ( { "feature-extraction": FalconModel, "text-classification": FalconForSequenceClassification, "text-generation": FalconForCausalLM, "question-answering": FalconForQuestionAnswering, "token-classification": FalconForTokenClassification, "zero-shot": FalconForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False def snake_case ( self : int ): """simple docstring""" __lowercase =FalconModelTester(self ) __lowercase =ConfigTester(self , config_class=__lowercase , hidden_size=37 ) def snake_case ( self : Dict ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self : Optional[Any] ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def snake_case ( self : Union[str, Any] ): """simple docstring""" __lowercase , *__lowercase =self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __lowercase =alibi self.model_tester.create_and_check_model(__lowercase , *__lowercase ) def snake_case ( self : str ): """simple docstring""" __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =3 __lowercase =input_dict['input_ids'] __lowercase =input_ids.ne(1 ).to(__lowercase ) __lowercase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __lowercase =FalconForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : Optional[Any] ): """simple docstring""" __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =3 __lowercase ='single_label_classification' __lowercase =input_dict['input_ids'] __lowercase =input_ids.ne(1 ).to(__lowercase ) __lowercase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __lowercase =FalconForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : int ): """simple docstring""" __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =input_dict['input_ids'] __lowercase =FalconForCausalLM(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , use_cache=__lowercase ) __lowercase =input_ids.shape[0] __lowercase =model._convert_to_rw_cache(result.past_key_values ) __lowercase =model._convert_cache_to_standard_format(__lowercase , __lowercase ) for layer in range(len(__lowercase ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def snake_case ( self : Union[str, Any] ): """simple docstring""" __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =3 __lowercase ='multi_label_classification' __lowercase =input_dict['input_ids'] __lowercase =input_ids.ne(1 ).to(__lowercase ) __lowercase =ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __lowercase =FalconForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : Tuple ): """simple docstring""" for model_class in self.all_generative_model_classes: __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(__lowercase , 'use_cache' ): return __lowercase =model_class(__lowercase ).to(__lowercase ) if "use_cache" not in inputs: __lowercase =True __lowercase =model(**__lowercase ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __lowercase =( getattr(__lowercase , 'decoder_layers' , __lowercase ) or getattr(__lowercase , 'num_decoder_layers' , __lowercase ) or config.num_hidden_layers ) __lowercase =getattr(__lowercase , 'num_kv_heads' , config.num_attention_heads ) __lowercase =getattr(__lowercase , 'd_model' , config.hidden_size ) __lowercase =embed_dim // num_attention_heads __lowercase =outputs['past_key_values'] self.assertEqual(len(__lowercase ) , __lowercase ) __lowercase , __lowercase =inputs['input_ids'].shape for i in range(__lowercase ): if config.new_decoder_architecture: __lowercase =config.num_attention_heads elif config.multi_query: __lowercase =1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class lowerCAmelCase ( unittest.TestCase ): @slow def snake_case ( self : List[str] ): """simple docstring""" __lowercase =AutoTokenizer.from_pretrained('Rocketknight1/falcon-rw-1b' ) __lowercase =FalconForCausalLM.from_pretrained('Rocketknight1/falcon-rw-1b' ) model.eval() model.to(__lowercase ) __lowercase =tokenizer('My favorite food is' , return_tensors='pt' ).to(__lowercase ) __lowercase =( 'My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday.' ) __lowercase =model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=19 ) __lowercase =tokenizer.batch_decode(__lowercase )[0] self.assertEqual(__lowercase , __lowercase ) @slow def snake_case ( self : Dict ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __lowercase =AutoTokenizer.from_pretrained(__lowercase ) __lowercase =FalconForCausalLM.from_pretrained(__lowercase ) model.eval() model.to(__lowercase ) __lowercase =tokenizer('My favorite food is' , return_tensors='pt' ).to(__lowercase ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=4 ) model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=4 ) model.generate(**__lowercase , num_beams=2 , max_new_tokens=4 ) @slow def snake_case ( self : Tuple ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __lowercase =AutoTokenizer.from_pretrained(__lowercase ) __lowercase =FalconForCausalLM.from_pretrained(__lowercase ) model.eval() model.to(device=__lowercase ) __lowercase =tokenizer('My favorite food is' , return_tensors='pt' ).to(__lowercase ) # Test results are the same with and without cache __lowercase =model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=20 , use_cache=__lowercase ) __lowercase =model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=20 , use_cache=__lowercase ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
141
1
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer A_ : Any = logging.get_logger(__name__) A_ : Dict = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} A_ : int = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } A_ : List[Any] = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } A_ : List[Any] = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } A_ : str = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } A_ : Optional[Any] = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } A_ : Any = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } A_ : List[Any] = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } A_ : int = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } A_ : List[Any] = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class _lowerCAmelCase( a_ ): """simple docstring""" a : Any =VOCAB_FILES_NAMES a : List[Any] =CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP a : Any =CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : List[Any] =CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION a : Optional[Any] =DPRContextEncoderTokenizer class _lowerCAmelCase( a_ ): """simple docstring""" a : int =VOCAB_FILES_NAMES a : List[Any] =QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP a : Union[str, Any] =QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : Tuple =QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION a : List[str] =DPRQuestionEncoderTokenizer A_ : List[Any] = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) A_ : List[Any] = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) A_ : Tuple = r'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(a_ ) class _lowerCAmelCase: """simple docstring""" def __call__( self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = False , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , **_lowerCamelCase , ): if titles is None and texts is None: return super().__call__( lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ , return_attention_mask=lowercase_ , **lowercase_ , ) elif titles is None or texts is None: UpperCamelCase_: List[Any] = titles if texts is None else texts return super().__call__( lowercase_ , lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ , return_attention_mask=lowercase_ , **lowercase_ , ) UpperCamelCase_: Optional[Any] = titles if not isinstance(lowercase_ , lowercase_ ) else [titles] UpperCamelCase_: List[Any] = texts if not isinstance(lowercase_ , lowercase_ ) else [texts] UpperCamelCase_: Optional[Any] = len(lowercase_ ) UpperCamelCase_: str = questions if not isinstance(lowercase_ , lowercase_ ) else [questions] * n_passages assert len(lowercase_ ) == len( lowercase_ ), f'''There should be as many titles than texts but got {len(lowercase_ )} titles and {len(lowercase_ )} texts.''' UpperCamelCase_: Any = super().__call__(lowercase_ , lowercase_ , padding=lowercase_ , truncation=lowercase_ )['input_ids'] UpperCamelCase_: int = super().__call__(lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ )['input_ids'] UpperCamelCase_: int = { 'input_ids': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowercase_ , lowercase_ ) ] } if return_attention_mask is not False: UpperCamelCase_: Optional[Any] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) UpperCamelCase_: Tuple = attention_mask return self.pad(lowercase_ , padding=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ ) def _a ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 1_6 , _lowerCamelCase = 6_4 , _lowerCamelCase = 4 , ): UpperCamelCase_: Tuple = reader_input['input_ids'] UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: str = reader_output[:3] UpperCamelCase_: Dict = len(lowercase_ ) UpperCamelCase_: Optional[int] = sorted(range(lowercase_ ) , reverse=lowercase_ , key=relevance_logits.__getitem__ ) UpperCamelCase_: Dict = [] for doc_id in sorted_docs: UpperCamelCase_: Tuple = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence UpperCamelCase_: Union[str, Any] = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: UpperCamelCase_: Optional[int] = sequence_ids.index(self.pad_token_id ) else: UpperCamelCase_: Union[str, Any] = len(lowercase_ ) UpperCamelCase_: Any = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=lowercase_ , top_spans=lowercase_ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=lowercase_ , start_index=lowercase_ , end_index=lowercase_ , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(lowercase_ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def _a ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): UpperCamelCase_: Optional[int] = [] for start_index, start_score in enumerate(lowercase_ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) UpperCamelCase_: List[Any] = sorted(lowercase_ , key=lambda _lowerCamelCase : x[1] , reverse=lowercase_ ) UpperCamelCase_: Union[str, Any] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f'''Wrong span indices: [{start_index}:{end_index}]''' UpperCamelCase_: Dict = end_index - start_index + 1 assert length <= max_answer_length, f'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(lowercase_ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(a_ ) class _lowerCAmelCase( a_ , a_ ): """simple docstring""" a : Any =VOCAB_FILES_NAMES a : Dict =READER_PRETRAINED_VOCAB_FILES_MAP a : Any =READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : List[Any] =READER_PRETRAINED_INIT_CONFIGURATION a : Tuple =['''input_ids''', '''attention_mask'''] a : Any =DPRReaderTokenizer
364
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input A_ : Any = 'Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine' def snake_case () -> Union[str, Any]: UpperCamelCase_: Tuple = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: UpperCamelCase_: List[str] = get_sagemaker_input() else: UpperCamelCase_: List[str] = get_cluster_input() return config def snake_case (UpperCAmelCase__=None ) -> Union[str, Any]: if subparsers is not None: UpperCamelCase_: List[Any] = subparsers.add_parser('config' , description=UpperCAmelCase__ ) else: UpperCamelCase_: List[Any] = argparse.ArgumentParser('Accelerate config command' , description=UpperCAmelCase__ ) parser.add_argument( '--config_file' , default=UpperCAmelCase__ , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=UpperCAmelCase__ ) return parser def snake_case (UpperCAmelCase__ ) -> List[Any]: UpperCamelCase_: Union[str, Any] = get_user_input() if args.config_file is not None: UpperCamelCase_: Tuple = args.config_file else: if not os.path.isdir(UpperCAmelCase__ ): os.makedirs(UpperCAmelCase__ ) UpperCamelCase_: Dict = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(UpperCAmelCase__ ) else: config.to_yaml_file(UpperCAmelCase__ ) print(F'''accelerate configuration saved at {config_file}''' ) def snake_case () -> str: UpperCamelCase_: Tuple = config_command_parser() UpperCamelCase_: int = parser.parse_args() config_command(UpperCAmelCase__ ) if __name__ == "__main__": main()
292
0
import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def A_ ( A__ ) -> Union[str, Any]: a__ : List[str] = split_dict._to_yaml_list() assert len(A__ ) == len(A__ ) a__ : List[Any] = SplitDict._from_yaml_list(A__ ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump a__ : Any = None # the split name of split_dict takes over the name of the split info object a__ : str = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=A__ ), SplitInfo(dataset_name='my_dataset' )] ) def A_ ( A__ ) -> Any: # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files a__ : Any = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
99
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __snake_case = None __snake_case = logging.get_logger(__name__) __snake_case = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} __snake_case = { '''vocab_file''': { '''google/bigbird-roberta-base''': '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model''', '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model''' ), }, '''tokenizer_file''': { '''google/bigbird-roberta-base''': ( '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json''' ), '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json''' ), }, } __snake_case = { '''google/bigbird-roberta-base''': 40_96, '''google/bigbird-roberta-large''': 40_96, '''google/bigbird-base-trivia-itc''': 40_96, } __snake_case = '''▁''' class __snake_case ( lowerCamelCase__ ): __lowerCamelCase : Dict = VOCAB_FILES_NAMES __lowerCamelCase : List[Any] = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase : List[str] = BigBirdTokenizer __lowerCamelCase : Any = ["""input_ids""", """attention_mask"""] __lowerCamelCase : List[int] = [] def __init__( self , snake_case__=None , snake_case__=None , snake_case__="<unk>" , snake_case__="<s>" , snake_case__="</s>" , snake_case__="<pad>" , snake_case__="[SEP]" , snake_case__="[MASK]" , snake_case__="[CLS]" , **snake_case__ , ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Any =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else bos_token UpperCAmelCase : Optional[int] =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else eos_token UpperCAmelCase : List[str] =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else unk_token UpperCAmelCase : Union[str, Any] =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else pad_token UpperCAmelCase : int =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else cls_token UpperCAmelCase : str =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : List[Any] =AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( snake_case__ , tokenizer_file=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , **snake_case__ , ) UpperCAmelCase : Tuple =vocab_file UpperCAmelCase : Optional[int] =False if not self.vocab_file else True def UpperCAmelCase__ ( self , snake_case__ , snake_case__ = None ) -> List[int]: '''simple docstring''' UpperCAmelCase : int =[self.sep_token_id] UpperCAmelCase : Optional[int] =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase__ ( self , snake_case__ , snake_case__ = None , snake_case__ = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1] + ([0] * len(snake_case__ )) + [1] def UpperCAmelCase__ ( self , snake_case__ , snake_case__ = None ) -> List[int]: '''simple docstring''' UpperCAmelCase : Optional[Any] =[self.sep_token_id] UpperCAmelCase : Optional[int] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self , snake_case__ , snake_case__ = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(snake_case__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase : Optional[int] =os.path.join( snake_case__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
348
0
"""simple docstring""" def a__ ( _SCREAMING_SNAKE_CASE = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" UpperCamelCase = set() # Replace all the whitespace in our sentence UpperCamelCase = input_str.replace(" " , "" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowerCamelCase_ ) == 26 def a__ ( _SCREAMING_SNAKE_CASE = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" UpperCamelCase = [False] * 26 for char in input_str: if char.islower(): UpperCamelCase = True elif char.isupper(): UpperCamelCase = True return all(lowerCamelCase_ ) def a__ ( _SCREAMING_SNAKE_CASE = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def a__ ( ): """simple docstring""" from timeit import timeit UpperCamelCase = '''from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest''' print(timeit("is_pangram()" , setup=lowerCamelCase_ ) ) print(timeit("is_pangram_faster()" , setup=lowerCamelCase_ ) ) print(timeit("is_pangram_fastest()" , setup=lowerCamelCase_ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
356
"""simple docstring""" import math def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = len(_SCREAMING_SNAKE_CASE ) UpperCamelCase = int(math.floor(math.sqrt(_SCREAMING_SNAKE_CASE ) ) ) UpperCamelCase = 0 while arr[min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) - 1] < x: UpperCamelCase = step step += int(math.floor(math.sqrt(_SCREAMING_SNAKE_CASE ) ) ) if prev >= n: return -1 while arr[prev] < x: UpperCamelCase = prev + 1 if prev == min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] lowerCAmelCase__ = int(input('''Enter the number to be searched:\n''')) lowerCAmelCase__ = jump_search(arr, x) if res == -1: print('''Number not found!''') else: print(f'''Number {x} is at index {res}''')
244
0
def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[str] = 0 # if input_string is "aba" than new_input_string become "a|b|a" lowerCamelCase : Union[str, Any] = """""" lowerCamelCase : str = """""" # append each character + "|" in new_string for range(0, length-1) for i in input_string[: len(_a ) - 1]: new_input_string += i + "|" # append last character new_input_string += input_string[-1] # we will store the starting and ending of previous furthest ending palindromic # substring lowerCamelCase : Tuple = 0, 0 # length[i] shows the length of palindromic substring with center i lowerCamelCase : Tuple = [1 for i in range(len(_a ) )] # for each character in new_string find corresponding palindromic string lowerCamelCase : Optional[Any] = 0 for j in range(len(_a ) ): lowerCamelCase : Tuple = 1 if j > r else min(length[l + r - j] // 2 , r - j + 1 ) while ( j - k >= 0 and j + k < len(_a ) and new_input_string[k + j] == new_input_string[j - k] ): k += 1 lowerCamelCase : Optional[Any] = 2 * k - 1 # does this string is ending after the previously explored end (that is r) ? # if yes the update the new r to the last index of this if j + k - 1 > r: lowerCamelCase : int = j - k + 1 # noqa: E741 lowerCamelCase : List[str] = j + k - 1 # update max_length and start position if max_length < length[j]: lowerCamelCase : List[str] = length[j] lowerCamelCase : List[str] = j # create that string lowerCamelCase : Optional[Any] = new_input_string[start - max_length // 2 : start + max_length // 2 + 1] for i in s: if i != "|": output_string += i return output_string if __name__ == "__main__": import doctest doctest.testmod()
283
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Dict = AutoencoderKL A__ : Optional[int] = "sample" A__ : Tuple = 1E-2 @property def A__ ( self: List[Any] ) -> Union[str, Any]: UpperCAmelCase_ : Tuple = 4 UpperCAmelCase_ : str = 3 UpperCAmelCase_ : Any = (32, 32) UpperCAmelCase_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase_ ) return {"sample": image} @property def A__ ( self: List[str] ) -> Tuple: return (3, 32, 32) @property def A__ ( self: Optional[Any] ) -> Any: return (3, 32, 32) def A__ ( self: Any ) -> Tuple: UpperCAmelCase_ : List[Any] = { """block_out_channels""": [32, 64], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 4, } UpperCAmelCase_ : int = self.dummy_input return init_dict, inputs_dict def A__ ( self: Optional[Any] ) -> int: pass def A__ ( self: str ) -> Any: pass @unittest.skipIf(torch_device == """mps""" ,"""Gradient checkpointing skipped on MPS""" ) def A__ ( self: Union[str, Any] ) -> Dict: # enable deterministic behavior for gradient checkpointing UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.prepare_init_args_and_inputs_for_common() UpperCAmelCase_ : List[Any] = self.model_class(**lowerCamelCase_ ) model.to(lowerCamelCase_ ) assert not model.is_gradient_checkpointing and model.training UpperCAmelCase_ : Optional[Any] = model(**lowerCamelCase_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() UpperCAmelCase_ : Any = torch.randn_like(lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing UpperCAmelCase_ : str = self.model_class(**lowerCamelCase_ ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase_ ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training UpperCAmelCase_ : Optional[int] = model_a(**lowerCamelCase_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() UpperCAmelCase_ : Dict = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1e-5 ) UpperCAmelCase_ : Dict = dict(model.named_parameters() ) UpperCAmelCase_ : Union[str, Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data ,named_params_a[name].grad.data ,atol=5e-5 ) ) def A__ ( self: Optional[Any] ) -> str: UpperCAmelCase_ , UpperCAmelCase_ : int = AutoencoderKL.from_pretrained("""fusing/autoencoder-kl-dummy""" ,output_loading_info=lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) self.assertEqual(len(loading_info["""missing_keys"""] ) ,0 ) model.to(lowerCamelCase_ ) UpperCAmelCase_ : Dict = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def A__ ( self: Optional[int] ) -> int: UpperCAmelCase_ : Dict = AutoencoderKL.from_pretrained("""fusing/autoencoder-kl-dummy""" ) UpperCAmelCase_ : Tuple = model.to(lowerCamelCase_ ) model.eval() if torch_device == "mps": UpperCAmelCase_ : Tuple = torch.manual_seed(0 ) else: UpperCAmelCase_ : Optional[int] = torch.Generator(device=lowerCamelCase_ ).manual_seed(0 ) UpperCAmelCase_ : str = torch.randn( 1 ,model.config.in_channels ,model.config.sample_size ,model.config.sample_size ,generator=torch.manual_seed(0 ) ,) UpperCAmelCase_ : int = image.to(lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : Dict = model(lowerCamelCase_ ,sample_posterior=lowerCamelCase_ ,generator=lowerCamelCase_ ).sample UpperCAmelCase_ : Optional[int] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": UpperCAmelCase_ : Tuple = torch.tensor( [ -4.0078e-01, -3.8323e-04, -1.2681e-01, -1.1462e-01, 2.0095e-01, 1.0893e-01, -8.8247e-02, -3.0361e-01, -9.8644e-03, ] ) elif torch_device == "cpu": UpperCAmelCase_ : List[str] = torch.tensor( [-0.1_3_5_2, 0.0_8_7_8, 0.0_4_1_9, -0.0_8_1_8, -0.1_0_6_9, 0.0_6_8_8, -0.1_4_5_8, -0.4_4_4_6, -0.0_0_2_6] ) else: UpperCAmelCase_ : List[str] = torch.tensor( [-0.2_4_2_1, 0.4_6_4_2, 0.2_5_0_7, -0.0_4_3_8, 0.0_6_8_2, 0.3_1_6_0, -0.2_0_1_8, -0.0_7_2_7, 0.2_4_8_5] ) self.assertTrue(torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,rtol=1e-2 ) ) @slow class _snake_case ( unittest.TestCase ): '''simple docstring''' def A__ ( self: Any ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Any ) -> Optional[Any]: return F'''gaussian_noise_s={seed}_shape={'_'.join([str(lowerCamelCase_ ) for s in shape] )}.npy''' def A__ ( self: Union[str, Any] ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self: List[str] ,lowerCamelCase_: Optional[int]=0 ,lowerCamelCase_: List[Any]=(4, 3, 512, 512) ,lowerCamelCase_: Optional[Any]=False ) -> Optional[int]: UpperCAmelCase_ : Tuple = torch.floataa if fpaa else torch.floataa UpperCAmelCase_ : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase_ ,lowerCamelCase_ ) ) ).to(lowerCamelCase_ ).to(lowerCamelCase_ ) return image def A__ ( self: List[Any] ,lowerCamelCase_: List[str]="CompVis/stable-diffusion-v1-4" ,lowerCamelCase_: Union[str, Any]=False ) -> Any: UpperCAmelCase_ : Optional[Any] = """fp16""" if fpaa else None UpperCAmelCase_ : str = torch.floataa if fpaa else torch.floataa UpperCAmelCase_ : int = AutoencoderKL.from_pretrained( lowerCamelCase_ ,subfolder="""vae""" ,torch_dtype=lowerCamelCase_ ,revision=lowerCamelCase_ ,) model.to(lowerCamelCase_ ).eval() return model def A__ ( self: Dict ,lowerCamelCase_: Union[str, Any]=0 ) -> Optional[int]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase_ ) return torch.Generator(device=lowerCamelCase_ ).manual_seed(lowerCamelCase_ ) @parameterized.expand( [ # fmt: off [33, [-0.1_6_0_3, 0.9_8_7_8, -0.0_4_9_5, -0.0_7_9_0, -0.2_7_0_9, 0.8_3_7_5, -0.2_0_6_0, -0.0_8_2_4], [-0.2_3_9_5, 0.0_0_9_8, 0.0_1_0_2, -0.0_7_0_9, -0.2_8_4_0, -0.0_2_7_4, -0.0_7_1_8, -0.1_8_2_4]], [47, [-0.2_3_7_6, 0.1_1_6_8, 0.1_3_3_2, -0.4_8_4_0, -0.2_5_0_8, -0.0_7_9_1, -0.0_4_9_3, -0.4_0_8_9], [0.0_3_5_0, 0.0_8_4_7, 0.0_4_6_7, 0.0_3_4_4, -0.0_8_4_2, -0.0_5_4_7, -0.0_6_3_3, -0.1_1_3_1]], # fmt: on ] ) def A__ ( self: List[Any] ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: str ,lowerCamelCase_: Dict ) -> Tuple: UpperCAmelCase_ : List[Any] = self.get_sd_vae_model() UpperCAmelCase_ : int = self.get_sd_image(lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = self.get_generator(lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ ,generator=lowerCamelCase_ ,sample_posterior=lowerCamelCase_ ).sample assert sample.shape == image.shape UpperCAmelCase_ : Optional[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() UpperCAmelCase_ : Tuple = torch.tensor(expected_slice_mps if torch_device == """mps""" else expected_slice ) assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=3e-3 ) @parameterized.expand( [ # fmt: off [33, [-0.0_5_1_3, 0.0_2_8_9, 1.3_7_9_9, 0.2_1_6_6, -0.2_5_7_3, -0.0_8_7_1, 0.5_1_0_3, -0.0_9_9_9]], [47, [-0.4_1_2_8, -0.1_3_2_0, -0.3_7_0_4, 0.1_9_6_5, -0.4_1_1_6, -0.2_3_3_2, -0.3_3_4_0, 0.2_2_4_7]], # fmt: on ] ) @require_torch_gpu def A__ ( self: Union[str, Any] ,lowerCamelCase_: Any ,lowerCamelCase_: List[str] ) -> Tuple: UpperCAmelCase_ : List[str] = self.get_sd_vae_model(fpaa=lowerCamelCase_ ) UpperCAmelCase_ : Any = self.get_sd_image(lowerCamelCase_ ,fpaa=lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = self.get_generator(lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ ,generator=lowerCamelCase_ ,sample_posterior=lowerCamelCase_ ).sample assert sample.shape == image.shape UpperCAmelCase_ : Tuple = sample[-1, -2:, :2, -2:].flatten().float().cpu() UpperCAmelCase_ : Optional[int] = torch.tensor(lowerCamelCase_ ) assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.1_6_0_9, 0.9_8_6_6, -0.0_4_8_7, -0.0_7_7_7, -0.2_7_1_6, 0.8_3_6_8, -0.2_0_5_5, -0.0_8_1_4], [-0.2_3_9_5, 0.0_0_9_8, 0.0_1_0_2, -0.0_7_0_9, -0.2_8_4_0, -0.0_2_7_4, -0.0_7_1_8, -0.1_8_2_4]], [47, [-0.2_3_7_7, 0.1_1_4_7, 0.1_3_3_3, -0.4_8_4_1, -0.2_5_0_6, -0.0_8_0_5, -0.0_4_9_1, -0.4_0_8_5], [0.0_3_5_0, 0.0_8_4_7, 0.0_4_6_7, 0.0_3_4_4, -0.0_8_4_2, -0.0_5_4_7, -0.0_6_3_3, -0.1_1_3_1]], # fmt: on ] ) def A__ ( self: Tuple ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: List[str] ) -> Dict: UpperCAmelCase_ : Optional[int] = self.get_sd_vae_model() UpperCAmelCase_ : Dict = self.get_sd_image(lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : str = model(lowerCamelCase_ ).sample assert sample.shape == image.shape UpperCAmelCase_ : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() UpperCAmelCase_ : Any = torch.tensor(expected_slice_mps if torch_device == """mps""" else expected_slice ) assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=3e-3 ) @parameterized.expand( [ # fmt: off [13, [-0.2_0_5_1, -0.1_8_0_3, -0.2_3_1_1, -0.2_1_1_4, -0.3_2_9_2, -0.3_5_7_4, -0.2_9_5_3, -0.3_3_2_3]], [37, [-0.2_6_3_2, -0.2_6_2_5, -0.2_1_9_9, -0.2_7_4_1, -0.4_5_3_9, -0.4_9_9_0, -0.3_7_2_0, -0.4_9_2_5]], # fmt: on ] ) @require_torch_gpu def A__ ( self: Optional[Any] ,lowerCamelCase_: Tuple ,lowerCamelCase_: str ) -> Optional[Any]: UpperCAmelCase_ : List[str] = self.get_sd_vae_model() UpperCAmelCase_ : Optional[int] = self.get_sd_image(lowerCamelCase_ ,shape=(3, 4, 64, 64) ) with torch.no_grad(): UpperCAmelCase_ : str = model.decode(lowerCamelCase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] UpperCAmelCase_ : Any = sample[-1, -2:, :2, -2:].flatten().cpu() UpperCAmelCase_ : Union[str, Any] = torch.tensor(lowerCamelCase_ ) assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=1e-3 ) @parameterized.expand( [ # fmt: off [27, [-0.0_3_6_9, 0.0_2_0_7, -0.0_7_7_6, -0.0_6_8_2, -0.1_7_4_7, -0.1_9_3_0, -0.1_4_6_5, -0.2_0_3_9]], [16, [-0.1_6_2_8, -0.2_1_3_4, -0.2_7_4_7, -0.2_6_4_2, -0.3_7_7_4, -0.4_4_0_4, -0.3_6_8_7, -0.4_2_7_7]], # fmt: on ] ) @require_torch_gpu def A__ ( self: str ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Any ) -> Optional[Any]: UpperCAmelCase_ : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = self.get_sd_image(lowerCamelCase_ ,shape=(3, 4, 64, 64) ,fpaa=lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : List[str] = model.decode(lowerCamelCase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] UpperCAmelCase_ : str = sample[-1, -2:, :2, -2:].flatten().float().cpu() UpperCAmelCase_ : str = torch.tensor(lowerCamelCase_ ) assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=5e-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() ,reason="""xformers is not required when using PyTorch 2.0.""" ) def A__ ( self: List[Any] ,lowerCamelCase_: Union[str, Any] ) -> int: UpperCAmelCase_ : Optional[Any] = self.get_sd_vae_model(fpaa=lowerCamelCase_ ) UpperCAmelCase_ : List[str] = self.get_sd_image(lowerCamelCase_ ,shape=(3, 4, 64, 64) ,fpaa=lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model.decode(lowerCamelCase_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): UpperCAmelCase_ : List[str] = model.decode(lowerCamelCase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=1e-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() ,reason="""xformers is not required when using PyTorch 2.0.""" ) def A__ ( self: Optional[Any] ,lowerCamelCase_: Dict ) -> Union[str, Any]: UpperCAmelCase_ : Tuple = self.get_sd_vae_model() UpperCAmelCase_ : Any = self.get_sd_image(lowerCamelCase_ ,shape=(3, 4, 64, 64) ) with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model.decode(lowerCamelCase_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model.decode(lowerCamelCase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.3_0_0_1, 0.0_9_1_8, -2.6_9_8_4, -3.9_7_2_0, -3.2_0_9_9, -5.0_3_5_3, 1.7_3_3_8, -0.2_0_6_5, 3.4_2_6_7]], [47, [-1.5_0_3_0, -4.3_8_7_1, -6.0_3_5_5, -9.1_1_5_7, -1.6_6_6_1, -2.7_8_5_3, 2.1_6_0_7, -5.0_8_2_3, 2.5_6_3_3]], # fmt: on ] ) def A__ ( self: Union[str, Any] ,lowerCamelCase_: Any ,lowerCamelCase_: Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase_ : Dict = self.get_sd_vae_model() UpperCAmelCase_ : Optional[Any] = self.get_sd_image(lowerCamelCase_ ) UpperCAmelCase_ : str = self.get_generator(lowerCamelCase_ ) with torch.no_grad(): UpperCAmelCase_ : int = model.encode(lowerCamelCase_ ).latent_dist UpperCAmelCase_ : Optional[Any] = dist.sample(generator=lowerCamelCase_ ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] UpperCAmelCase_ : Tuple = sample[0, -1, -3:, -3:].flatten().cpu() UpperCAmelCase_ : Optional[Any] = torch.tensor(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = 3e-3 if torch_device != """mps""" else 1e-2 assert torch_all_close(lowerCamelCase_ ,lowerCamelCase_ ,atol=lowerCamelCase_ )
345
0
'''simple docstring''' import math import sys def __A ( lowerCAmelCase_ ): _UpperCAmelCase : Union[str, Any] = """""" try: with open(lowerCAmelCase_ , """rb""" ) as binary_file: _UpperCAmelCase : int = binary_file.read() for dat in data: _UpperCAmelCase : Any = f"{dat:08b}" result += curr_byte return result except OSError: print("""File not accessible""" ) sys.exit() def __A ( lowerCAmelCase_ ): _UpperCAmelCase : Optional[Any] = {"""0""": """0""", """1""": """1"""} _UpperCAmelCase , _UpperCAmelCase : List[str] = """""", """""" _UpperCAmelCase : Any = len(lowerCAmelCase_ ) for i in range(len(lowerCAmelCase_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue _UpperCAmelCase : int = lexicon[curr_string] result += last_match_id _UpperCAmelCase : Optional[Any] = last_match_id + """0""" if math.loga(lowerCAmelCase_ ).is_integer(): _UpperCAmelCase : List[Any] = {} for curr_key in list(lowerCAmelCase_ ): _UpperCAmelCase : Any = lexicon.pop(lowerCAmelCase_ ) _UpperCAmelCase : Optional[int] = new_lex _UpperCAmelCase : Any = last_match_id + """1""" index += 1 _UpperCAmelCase : Union[str, Any] = """""" return result def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): _UpperCAmelCase : Dict = 8 try: with open(lowerCAmelCase_ , """wb""" ) as opened_file: _UpperCAmelCase : Union[str, Any] = [ to_write[i : i + byte_length] for i in range(0 , len(lowerCAmelCase_ ) , lowerCAmelCase_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append("""10000000""" ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(lowerCAmelCase_ , 2 ).to_bytes(1 , byteorder="""big""" ) ) except OSError: print("""File not accessible""" ) sys.exit() def __A ( lowerCAmelCase_ ): _UpperCAmelCase : Any = 0 for letter in data_bits: if letter == "1": break counter += 1 _UpperCAmelCase : Optional[Any] = data_bits[counter:] _UpperCAmelCase : Optional[Any] = data_bits[counter + 1 :] return data_bits def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): _UpperCAmelCase : Tuple = read_file_binary(lowerCAmelCase_ ) _UpperCAmelCase : str = remove_prefix(lowerCAmelCase_ ) _UpperCAmelCase : Any = decompress_data(lowerCAmelCase_ ) write_file_binary(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
170
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : int = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[Any] = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys lowerCAmelCase_ : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
170
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ : int = { "configuration_bert": ["BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BertConfig", "BertOnnxConfig"], "tokenization_bert": ["BasicTokenizer", "BertTokenizer", "WordpieceTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[Any] = ["BertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[Any] = [ "BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "BertForMaskedLM", "BertForMultipleChoice", "BertForNextSentencePrediction", "BertForPreTraining", "BertForQuestionAnswering", "BertForSequenceClassification", "BertForTokenClassification", "BertLayer", "BertLMHeadModel", "BertModel", "BertPreTrainedModel", "load_tf_weights_in_bert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[Any] = [ "TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFBertEmbeddings", "TFBertForMaskedLM", "TFBertForMultipleChoice", "TFBertForNextSentencePrediction", "TFBertForPreTraining", "TFBertForQuestionAnswering", "TFBertForSequenceClassification", "TFBertForTokenClassification", "TFBertLMHeadModel", "TFBertMainLayer", "TFBertModel", "TFBertPreTrainedModel", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Any = ["TFBertTokenizer"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Union[str, Any] = [ "FlaxBertForCausalLM", "FlaxBertForMaskedLM", "FlaxBertForMultipleChoice", "FlaxBertForNextSentencePrediction", "FlaxBertForPreTraining", "FlaxBertForQuestionAnswering", "FlaxBertForSequenceClassification", "FlaxBertForTokenClassification", "FlaxBertModel", "FlaxBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys A_ : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
165
"""simple docstring""" import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def A ( snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' return params[f"""{prefix}/{prefix}/relpos_bias/rel_embedding"""][:, i, :] def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__="attention" ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = SCREAMING_SNAKE_CASE__ = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/key/kernel"""][:, i, :, :] ) SCREAMING_SNAKE_CASE__ = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) SCREAMING_SNAKE_CASE__ = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/out/kernel"""][:, i, :, :] ) SCREAMING_SNAKE_CASE__ = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) SCREAMING_SNAKE_CASE__ = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/query/kernel"""][:, i, :, :] ) SCREAMING_SNAKE_CASE__ = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) SCREAMING_SNAKE_CASE__ = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/value/kernel"""][:, i, :, :] ) SCREAMING_SNAKE_CASE__ = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__=False ): '''simple docstring''' if split_mlp_wi: SCREAMING_SNAKE_CASE__ = params[f"""{prefix}/{prefix}/mlp/wi_0/kernel"""][:, i, :] SCREAMING_SNAKE_CASE__ = params[f"""{prefix}/{prefix}/mlp/wi_1/kernel"""][:, i, :] SCREAMING_SNAKE_CASE__ = (wi_a, wi_a) else: SCREAMING_SNAKE_CASE__ = params[f"""{prefix}/{prefix}/mlp/wi/kernel"""][:, i, :] SCREAMING_SNAKE_CASE__ = params[f"""{prefix}/{prefix}/mlp/wo/kernel"""][:, i, :] return wi, wo def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' return params[f"""{prefix}/{prefix}/{layer_name}/scale"""][:, i] def A ( snake_case__ , *, snake_case__ , snake_case__ , snake_case__ = False ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = traverse_util.flatten_dict(variables["""target"""] ) SCREAMING_SNAKE_CASE__ = {"""/""".join(snake_case__ ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi SCREAMING_SNAKE_CASE__ = """encoder/encoder/mlp/wi_0/kernel""" in old print("""Split MLP:""" , snake_case__ ) SCREAMING_SNAKE_CASE__ = collections.OrderedDict() # Shared embeddings. SCREAMING_SNAKE_CASE__ = old["""token_embedder/embedding"""] # Encoder. for i in range(snake_case__ ): # Block i, layer 0 (Self Attention). SCREAMING_SNAKE_CASE__ = tax_layer_norm_lookup(snake_case__ , snake_case__ , """encoder""" , """pre_attention_layer_norm""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = tax_attention_lookup(snake_case__ , snake_case__ , """encoder""" , """attention""" ) SCREAMING_SNAKE_CASE__ = layer_norm SCREAMING_SNAKE_CASE__ = k.T SCREAMING_SNAKE_CASE__ = o.T SCREAMING_SNAKE_CASE__ = q.T SCREAMING_SNAKE_CASE__ = v.T # Block i, layer 1 (MLP). SCREAMING_SNAKE_CASE__ = tax_layer_norm_lookup(snake_case__ , snake_case__ , """encoder""" , """pre_mlp_layer_norm""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = tax_mlp_lookup(snake_case__ , snake_case__ , """encoder""" , snake_case__ ) SCREAMING_SNAKE_CASE__ = layer_norm if split_mlp_wi: SCREAMING_SNAKE_CASE__ = wi[0].T SCREAMING_SNAKE_CASE__ = wi[1].T else: SCREAMING_SNAKE_CASE__ = wi.T SCREAMING_SNAKE_CASE__ = wo.T if scalable_attention: # convert the rel_embedding of each layer SCREAMING_SNAKE_CASE__ = tax_relpos_bias_lookup( snake_case__ , snake_case__ , """encoder""" ).T SCREAMING_SNAKE_CASE__ = old["""encoder/encoder_norm/scale"""] if not scalable_attention: SCREAMING_SNAKE_CASE__ = tax_relpos_bias_lookup( snake_case__ , 0 , """encoder""" ).T SCREAMING_SNAKE_CASE__ = tax_relpos_bias_lookup( snake_case__ , 0 , """decoder""" ).T if not is_encoder_only: # Decoder. for i in range(snake_case__ ): # Block i, layer 0 (Self Attention). SCREAMING_SNAKE_CASE__ = tax_layer_norm_lookup(snake_case__ , snake_case__ , """decoder""" , """pre_self_attention_layer_norm""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = tax_attention_lookup(snake_case__ , snake_case__ , """decoder""" , """self_attention""" ) SCREAMING_SNAKE_CASE__ = layer_norm SCREAMING_SNAKE_CASE__ = k.T SCREAMING_SNAKE_CASE__ = o.T SCREAMING_SNAKE_CASE__ = q.T SCREAMING_SNAKE_CASE__ = v.T # Block i, layer 1 (Cross Attention). SCREAMING_SNAKE_CASE__ = tax_layer_norm_lookup(snake_case__ , snake_case__ , """decoder""" , """pre_cross_attention_layer_norm""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = tax_attention_lookup(snake_case__ , snake_case__ , """decoder""" , """encoder_decoder_attention""" ) SCREAMING_SNAKE_CASE__ = layer_norm SCREAMING_SNAKE_CASE__ = k.T SCREAMING_SNAKE_CASE__ = o.T SCREAMING_SNAKE_CASE__ = q.T SCREAMING_SNAKE_CASE__ = v.T # Block i, layer 2 (MLP). SCREAMING_SNAKE_CASE__ = tax_layer_norm_lookup(snake_case__ , snake_case__ , """decoder""" , """pre_mlp_layer_norm""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = tax_mlp_lookup(snake_case__ , snake_case__ , """decoder""" , snake_case__ ) SCREAMING_SNAKE_CASE__ = layer_norm if split_mlp_wi: SCREAMING_SNAKE_CASE__ = wi[0].T SCREAMING_SNAKE_CASE__ = wi[1].T else: SCREAMING_SNAKE_CASE__ = wi.T SCREAMING_SNAKE_CASE__ = wo.T if scalable_attention: # convert the rel_embedding of each layer SCREAMING_SNAKE_CASE__ = tax_relpos_bias_lookup(snake_case__ , snake_case__ , """decoder""" ).T SCREAMING_SNAKE_CASE__ = old["""decoder/decoder_norm/scale"""] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: SCREAMING_SNAKE_CASE__ = old["""decoder/logits_dense/kernel"""].T return new def A ( snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: SCREAMING_SNAKE_CASE__ = state_dict["""shared.weight"""] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: SCREAMING_SNAKE_CASE__ = state_dict["""shared.weight"""] if "lm_head.weight" not in state_dict: # For old 1.0 models. print("""Using shared word embeddings as lm_head.""" ) SCREAMING_SNAKE_CASE__ = state_dict["""shared.weight"""] return state_dict def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = checkpoints.load_tax_checkpoint(snake_case__ ) SCREAMING_SNAKE_CASE__ = convert_tax_to_pytorch( snake_case__ , num_layers=config.num_layers , is_encoder_only=snake_case__ , scalable_attention=snake_case__ ) SCREAMING_SNAKE_CASE__ = make_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ , strict=snake_case__ ) def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = False , snake_case__ = False , ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = MTaConfig.from_json_file(snake_case__ ) print(f"""Building PyTorch model from configuration: {config}""" ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: SCREAMING_SNAKE_CASE__ = UMTaEncoderModel(snake_case__ ) else: SCREAMING_SNAKE_CASE__ = UMTaForConditionalGeneration(snake_case__ ) # Load weights from tf checkpoint load_tax_weights_in_ta(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(snake_case__ ) # Verify that we can load the checkpoint. model.from_pretrained(snake_case__ ) print("""Done""" ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser(description="Converts a native T5X checkpoint into a PyTorch checkpoint.") # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path to the T5X checkpoint." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--is_encoder_only", action="store_true", help="Check if the model is encoder-decoder model", default=False ) parser.add_argument( "--scalable_attention", action="store_true", help="Whether the model uses scaled attention (umt5 model)", default=False, ) A_ : int = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
165
1
import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase , A__ ): def __lowerCamelCase ( self ): lowercase : Optional[int] = load_tool('''text-to-speech''' ) self.tool.setup() def __lowerCamelCase ( self ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) lowercase : Dict = self.tool('''hey''' ) lowercase : Tuple = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) ) def __lowerCamelCase ( self ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) lowercase : int = self.tool('''hey''' ) lowercase : str = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) )
173
from __future__ import annotations from math import ceil, floor, sqrt def __lowercase ( _UpperCamelCase = 2000000 ) ->int: """simple docstring""" lowercase : list[int] = [0] lowercase : int for idx in range(1, ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target lowercase : int = 0 # the area corresponding to the grid that gives the product closest to target lowercase : int = 0 # an estimate of b, using the quadratic formula lowercase : float # the largest integer less than b_estimate lowercase : int # the largest integer less than b_estimate lowercase : int # the triangle number corresponding to b_floor lowercase : int # the triangle number corresponding to b_ceil lowercase : int for idx_a, triangle_a in enumerate(triangle_numbers[1:], 1 ): lowercase : List[str] = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 lowercase : str = floor(_UpperCamelCase ) lowercase : int = ceil(_UpperCamelCase ) lowercase : str = triangle_numbers[b_floor] lowercase : str = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): lowercase : Optional[int] = triangle_b_first_guess * triangle_a lowercase : Tuple = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): lowercase : Dict = triangle_b_second_guess * triangle_a lowercase : Any = idx_a * b_ceil return area if __name__ == "__main__": print(F'''{solution() = }''')
173
1
"""simple docstring""" import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[str] ): __lowercase = logging.get_logger() # the current default level is logging.WARNING __lowercase = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel(), logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel(), logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel(), logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel(), logging.get_verbosity() ) # restore to the original level logging.set_verbosity(UpperCAmelCase__ ) def _lowercase ( self : str ): __lowercase = logging.get_verbosity() __lowercase = logging.get_logger("transformers.models.bart.tokenization_bart" ) __lowercase = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(UpperCAmelCase__ ) as cl: logger.warning(UpperCAmelCase__ ) self.assertEqual(cl.out, msg + "\n" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(UpperCAmelCase__ ) as cl: logger.warning(UpperCAmelCase__ ) self.assertEqual(cl.out, "" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(UpperCAmelCase__ ) as cl: logger.warning(UpperCAmelCase__ ) self.assertEqual(cl.out, msg + "\n" ) # restore to the original level logging.set_verbosity(UpperCAmelCase__ ) @mockenv(TRANSFORMERS_VERBOSITY="error" ) def _lowercase ( self : Any ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() # this action activates the env var __lowercase = logging.get_logger("transformers.models.bart.tokenization_bart" ) __lowercase = os.getenv("TRANSFORMERS_VERBOSITY", UpperCAmelCase__ ) __lowercase = logging.log_levels[env_level_str] __lowercase = logging.get_verbosity() self.assertEqual( UpperCAmelCase__, UpperCAmelCase__, F"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""", ) # restore to the original level __lowercase = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error" ) def _lowercase ( self : str ): # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() __lowercase = logging.logging.getLogger() with CaptureLogger(UpperCAmelCase__ ) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart" ) self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error", cl.out ) # no need to restore as nothing was changed def _lowercase ( self : Union[str, Any] ): # testing `logger.warning_advice()` transformers.utils.logging._reset_library_root_logger() __lowercase = logging.get_logger("transformers.models.bart.tokenization_bart" ) __lowercase = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1" ): # nothing should be logged as env var disables this method with CaptureLogger(UpperCAmelCase__ ) as cl: logger.warning_advice(UpperCAmelCase__ ) self.assertEqual(cl.out, "" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(UpperCAmelCase__ ) as cl: logger.warning_advice(UpperCAmelCase__ ) self.assertEqual(cl.out, msg + "\n" ) def _A ( ) -> int: '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
17
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" ,"False" ) ) is not True ,reason="Skipping test because should only be run when releasing minor transformers version" ,) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue_model_parallelism.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_6_0_0, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_6_0_0, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, ] ) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Optional[int] ): if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split(), encoding="utf-8", check=UpperCAmelCase__, ) assert hasattr(self, "env" ) def _lowercase ( self : str, UpperCAmelCase__ : List[Any] ): # configuration for running training on smdistributed Model Parallel __lowercase = { "enabled": True, "processes_per_host": 8, } __lowercase = { "enabled": True, "parameters": { "microbatches": 4, "placement_strategy": "spread", "pipeline": "interleaved", "optimize": "speed", "partitions": 4, "ddp": True, }, } __lowercase = {"smdistributed": {"modelparallel": smp_options}, "mpi": mpi_options} __lowercase = "trainer" if self.script == "run_glue.py" else "smtrainer" # creates estimator return HuggingFace( entry_point=self.script, source_dir=self.env.test_path, role=self.env.role, image_uri=self.env.image_uri, base_job_name=F"""{self.env.base_job_name}-{instance_count}-smp-{name_extension}""", instance_count=UpperCAmelCase__, instance_type=self.instance_type, debugger_hook_config=UpperCAmelCase__, hyperparameters={ **self.env.hyperparameters, "model_name_or_path": self.model_name_or_path, "max_steps": 5_0_0, }, metric_definitions=self.env.metric_definitions, distribution=UpperCAmelCase__, py_version="py36", ) def _lowercase ( self : Tuple, UpperCAmelCase__ : int ): TrainingJobAnalytics(UpperCAmelCase__ ).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(1,)] ) def _lowercase ( self : str, UpperCAmelCase__ : Union[str, Any] ): # create estimator __lowercase = self.create_estimator(UpperCAmelCase__ ) # run training estimator.fit() # result dataframe __lowercase = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __lowercase = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) __lowercase = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowercase = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds", 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""", "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss}, UpperCAmelCase__ )
17
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case_ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase ): A_ : Optional[Any] = 'maskformer-swin' A_ : Tuple = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__(self : List[Any] , a__ : Optional[Any]=224 , a__ : List[str]=4 , a__ : Dict=3 , a__ : str=96 , a__ : str=[2, 2, 6, 2] , a__ : Dict=[3, 6, 12, 24] , a__ : Any=7 , a__ : List[Any]=4.0 , a__ : List[Any]=True , a__ : Optional[Any]=0.0 , a__ : Union[str, Any]=0.0 , a__ : Optional[Any]=0.1 , a__ : Any="gelu" , a__ : str=False , a__ : Optional[Any]=0.0_2 , a__ : Union[str, Any]=1E-5 , a__ : Any=None , a__ : int=None , **a__ : Dict , ): """simple docstring""" super().__init__(**a__ ) __snake_case = image_size __snake_case = patch_size __snake_case = num_channels __snake_case = embed_dim __snake_case = depths __snake_case = len(a__ ) __snake_case = num_heads __snake_case = window_size __snake_case = mlp_ratio __snake_case = qkv_bias __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = drop_path_rate __snake_case = hidden_act __snake_case = use_absolute_embeddings __snake_case = layer_norm_eps __snake_case = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __snake_case = int(embed_dim * 2 ** (len(a__ ) - 1) ) __snake_case = ['''stem'''] + [f"""stage{idx}""" for idx in range(1 , len(a__ ) + 1 )] __snake_case , __snake_case = get_aligned_output_features_output_indices( out_features=a__ , out_indices=a__ , stage_names=self.stage_names )
238
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) snake_case_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = ['NllbTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = ['NllbTokenizerFast'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys snake_case_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
238
1
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : int = logging.get_logger(__name__) snake_case_ : str = {} class __snake_case ( a ): UpperCAmelCase__ : str = '''llama''' UpperCAmelCase__ : Dict = ['''past_key_values'''] def __init__( self : str , _snake_case : List[str]=32000 , _snake_case : int=4096 , _snake_case : List[str]=11008 , _snake_case : Optional[int]=32 , _snake_case : List[Any]=32 , _snake_case : Tuple=None , _snake_case : int="silu" , _snake_case : List[Any]=2048 , _snake_case : List[str]=0.0_2 , _snake_case : Any=1e-6 , _snake_case : List[str]=True , _snake_case : Optional[Any]=0 , _snake_case : Dict=1 , _snake_case : List[Any]=2 , _snake_case : str=1 , _snake_case : Union[str, Any]=False , _snake_case : str=None , **_snake_case : List[Any] , ): """simple docstring""" UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads # for backward compatibility if num_key_value_heads is None: UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = num_key_value_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = initializer_range UpperCAmelCase_ = rms_norm_eps UpperCAmelCase_ = pretraining_tp UpperCAmelCase_ = use_cache UpperCAmelCase_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , tie_word_embeddings=_snake_case , **_snake_case , ) def lowerCamelCase ( self : Optional[int]): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _snake_case) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""") UpperCAmelCase_ = self.rope_scaling.get('''type''' , _snake_case) UpperCAmelCase_ = self.rope_scaling.get('''factor''' , _snake_case) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""") if rope_scaling_factor is None or not isinstance(_snake_case , _snake_case) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""")
51
import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __snake_case ( a ): UpperCAmelCase__ : Dict = ['''image_processor''', '''tokenizer'''] UpperCAmelCase__ : Dict = '''FlavaImageProcessor''' UpperCAmelCase__ : Dict = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self : Union[str, Any] , _snake_case : List[str]=None , _snake_case : str=None , **_snake_case : int): """simple docstring""" UpperCAmelCase_ = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) UpperCAmelCase_ = kwargs.pop('''feature_extractor''') UpperCAmelCase_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(_snake_case , _snake_case) UpperCAmelCase_ = self.image_processor def __call__( self : List[Any] , _snake_case : Optional[ImageInput] = None , _snake_case : Optional[Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]]] = None , _snake_case : bool = True , _snake_case : Union[bool, str, PaddingStrategy] = False , _snake_case : Union[bool, str, TruncationStrategy] = False , _snake_case : Optional[int] = None , _snake_case : int = 0 , _snake_case : Optional[int] = None , _snake_case : Optional[bool] = None , _snake_case : Optional[bool] = None , _snake_case : Optional[bool] = None , _snake_case : Optional[bool] = None , _snake_case : bool = False , _snake_case : bool = False , _snake_case : bool = False , _snake_case : bool = False , _snake_case : bool = True , _snake_case : Optional[Union[str, TensorType]] = None , **_snake_case : Any , ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''') if text is not None: UpperCAmelCase_ = self.tokenizer( text=_snake_case , add_special_tokens=_snake_case , padding=_snake_case , truncation=_snake_case , max_length=_snake_case , stride=_snake_case , pad_to_multiple_of=_snake_case , return_token_type_ids=_snake_case , return_attention_mask=_snake_case , return_overflowing_tokens=_snake_case , return_special_tokens_mask=_snake_case , return_offsets_mapping=_snake_case , return_length=_snake_case , verbose=_snake_case , return_tensors=_snake_case , **_snake_case , ) if images is not None: UpperCAmelCase_ = self.image_processor( _snake_case , return_image_mask=_snake_case , return_codebook_pixels=_snake_case , return_tensors=_snake_case , **_snake_case , ) if text is not None and images is not None: encoding.update(_snake_case) return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case) , tensor_type=_snake_case) def lowerCamelCase ( self : Any , *_snake_case : Optional[Any] , **_snake_case : int): """simple docstring""" return self.tokenizer.batch_decode(*_snake_case , **_snake_case) def lowerCamelCase ( self : Optional[int] , *_snake_case : int , **_snake_case : Dict): """simple docstring""" return self.tokenizer.decode(*_snake_case , **_snake_case) @property def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = self.tokenizer.model_input_names UpperCAmelCase_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) @property def lowerCamelCase ( self : str): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _snake_case , ) return self.image_processor_class @property def lowerCamelCase ( self : Any): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _snake_case , ) return self.image_processor
51
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = "timm_backbone" def __init__( self : str , _lowercase : List[str]=None , _lowercase : Optional[int]=3 , _lowercase : List[Any]=True , _lowercase : str=True , _lowercase : Optional[Any]=None , **_lowercase : Dict , ): """simple docstring""" super().__init__(**_lowercase ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
204
from __future__ import annotations __lowerCamelCase : Dict = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __lowerCamelCase : Union[str, Any] = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[float] ) -> list[float]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = len(__UpperCamelCase ) for i in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE__ = -1 for j in range(i + 1 , __UpperCamelCase ): if arr[i] < arr[j]: SCREAMING_SNAKE_CASE__ = arr[j] break result.append(__UpperCamelCase ) return result def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[float] ) -> list[float]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] for i, outer in enumerate(__UpperCamelCase ): SCREAMING_SNAKE_CASE__ = -1 for inner in arr[i + 1 :]: if outer < inner: SCREAMING_SNAKE_CASE__ = inner break result.append(__UpperCamelCase ) return result def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[float] ) -> list[float]: """simple docstring""" SCREAMING_SNAKE_CASE__ = len(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [-1] * arr_size for index in reversed(range(__UpperCamelCase ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: SCREAMING_SNAKE_CASE__ = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCamelCase : List[Any] = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
204
1
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class snake_case__( unittest.TestCase ): '''simple docstring''' def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) lowerCAmelCase_ : Optional[Any] = Vector() def lowercase_ ( self ) -> None: lowerCAmelCase_ : List[str] = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(__lowercase ) , '''(0,0,0,0,0,1)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : int = Vector([1, 2, 3, 4] ) self.assertEqual(len(__lowercase ) , 4 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[Any] = Vector([1, 2] ) lowerCAmelCase_ : Dict = Vector([1, 2, 3, 4, 5] ) lowerCAmelCase_ : Tuple = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) lowerCAmelCase_ : Dict = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.2_36 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.4_16 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.6_16 , 3 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Vector([1, 2, 3] ) lowerCAmelCase_ : Optional[Any] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[Any] = Vector([1, 2, 3] ) lowerCAmelCase_ : int = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Any = Vector([1, 2, 3] ) lowerCAmelCase_ : Union[str, Any] = Vector([2, -1, 4] ) # for test of dot product lowerCAmelCase_ : List[str] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def lowercase_ ( self ) -> None: self.assertEqual(str(zero_vector(1_0 ) ).count('''0''' ) , 1_0 ) def lowercase_ ( self ) -> None: self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Dict = Vector([1, 2, 3] ) lowerCAmelCase_ : Tuple = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , __lowercase , __lowercase ) ) , '''(3,4,7)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 0, 0, 0, 0, 0] ) lowerCAmelCase_ : List[str] = x.copy() self.assertEqual(str(__lowercase ) , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(__lowercase ) , '''(0,1,0)''' ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : int = [[-3, -1_4, -1_0], [-5, -1_0, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(__lowercase , __lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : str = [[-3, 1_4, -1_0], [5, -1_0, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(__lowercase , __lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Any = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) lowerCAmelCase_ : Dict = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(__lowercase ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def lowercase_ ( self ) -> None: lowerCAmelCase_ : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) lowerCAmelCase_ : str = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def lowercase_ ( self ) -> None: self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
262
from typing import Union import fire import torch from tqdm import tqdm def lowerCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ = "cpu" , lowerCAmelCase_ = None )-> None: lowerCAmelCase_ : str = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCAmelCase_ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowerCAmelCase_ : int = v.half() if save_path is None: # overwrite src_path lowerCAmelCase_ : Tuple = src_path torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": fire.Fire(convert)
262
1
_lowerCAmelCase : Dict = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def UpperCamelCase_( _snake_case : List[Any] , _snake_case : Tuple , _snake_case : Any , _snake_case : Dict ): """simple docstring""" __a =[False] * len(_snake_case ) __a =[s] __a =True while queue: __a =queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_snake_case ) __a =True __a =u return visited[t] def UpperCamelCase_( _snake_case : int , _snake_case : Optional[int] , _snake_case : Union[str, Any] ): """simple docstring""" __a =[-1] * (len(_snake_case )) __a =0 __a =[] __a =[i[:] for i in graph] # Record original cut, copy. while bfs(_snake_case , _snake_case , _snake_case , _snake_case ): __a =float('Inf' ) __a =sink while s != source: # Find the minimum value in select path __a =min(_snake_case , graph[parent[s]][s] ) __a =parent[s] max_flow += path_flow __a =sink while v != source: __a =parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow __a =parent[v] for i in range(len(_snake_case ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
308
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _lowerCAmelCase : Optional[int] = logging.getLogger(__name__) _lowerCAmelCase : Any = "pytorch_model.bin" @dataclasses.dataclass class __magic_name__ : SCREAMING_SNAKE_CASE = dataclasses.field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models.'} ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co.'} , ) @dataclasses.dataclass class __magic_name__ : SCREAMING_SNAKE_CASE = dataclasses.field(metadata={'help': 'A csv or a json file containing the training data.'} ) SCREAMING_SNAKE_CASE = dataclasses.field(metadata={'help': 'A csv or a json file containing the data to predict on.'} ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'A csv or a json file containing the validation data.'} ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'The name of the task to train on.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'The list of labels for the task.'} ) @dataclasses.dataclass class __magic_name__ : SCREAMING_SNAKE_CASE = dataclasses.field( metadata={'help': 'The output directory where the model predictions and checkpoints will be written.'} ) SCREAMING_SNAKE_CASE = dataclasses.field( default='accuracy' , metadata={'help': 'The evaluation metric used for the task.'} ) SCREAMING_SNAKE_CASE = dataclasses.field( default='no' , metadata={ 'help': 'The evaluation strategy to adopt during training. Possible values are: ["no", "step", "epoch]' } , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=1_0 , metadata={'help': 'Number of evaluation calls with no improvement after which training will be stopped.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=0.0 , metadata={ 'help': 'How much the specified evaluation metric must improve to satisfy early stopping conditions.' } , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'Whether to filter the pseudo-labeled data based on the confidence score.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'Whether to filter the pseudo-labeled data based on the validation performance.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'Whether to fine-tune on labeled data after pseudo training.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=0.0 , metadata={'help': 'Confidence threshold for pseudo-labeled data filtering.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=1_0_0 , metadata={'help': 'Number of evaluation calls with no improvement after which training will be stopped.'} , ) SCREAMING_SNAKE_CASE = dataclasses.field( default=lowerCAmelCase_ , metadata={'help': 'Random seed for initialization.'} , ) def UpperCamelCase_( _snake_case : int , _snake_case : str , _snake_case : Optional[int] , _snake_case : str , _snake_case : Union[str, Any] , _snake_case : List[Any] ): """simple docstring""" __a =datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: __a =dataset.filter(lambda _snake_case : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 __a =int(eval_result * len(_snake_case ) ) print(_snake_case ) __a =dataset.sort('probability' , reverse=_snake_case ) __a =dataset.select(range(_snake_case ) ) __a =dataset.remove_columns(['label', 'probability'] ) __a =dataset.rename_column('prediction' , 'label' ) __a =dataset.map(lambda _snake_case : {"label": idalabel[example["label"]]} ) __a =dataset.shuffle(seed=args.seed ) __a =os.path.join(_snake_case , F'train_pseudo.{args.data_file_extension}' ) if args.data_file_extension == "csv": dataset.to_csv(_snake_case , index=_snake_case ) else: dataset.to_json(_snake_case ) def UpperCamelCase_( _snake_case : List[Any] , _snake_case : str , _snake_case : int , _snake_case : Optional[int] , **_snake_case : List[str] ): """simple docstring""" __a =Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() __a =STModelArguments(model_name_or_path=_snake_case ) __a =STDataArguments(train_file=_snake_case , infer_file=_snake_case ) __a =STTrainingArguments(output_dir=_snake_case ) __a =argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(_snake_case ).items(): setattr(_snake_case , _snake_case , _snake_case ) for key, value in kwargs.items(): if hasattr(_snake_case , _snake_case ): setattr(_snake_case , _snake_case , _snake_case ) # Sanity checks __a ={} __a =None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None __a =args.train_file __a =args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None __a =args.eval_file for key in data_files: __a =data_files[key].split('.' )[-1] assert extension in ["csv", "json"], F'`{key}_file` should be a csv or a json file.' if args.data_file_extension is None: __a =extension else: assert extension == args.data_file_extension, F'`{key}_file` should be a {args.data_file_extension} file`.' assert ( args.eval_metric in datasets.list_metrics() ), F'{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info('Creating the initial data directory for self-training...' ) __a =F'{args.output_dir}/self-train_iter-{{}}'.format __a =data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=_snake_case ) os.makedirs(_snake_case , exist_ok=_snake_case ) accelerator.wait_for_everyone() __a =None __a =None __a =0 __a =False # Show the progress bar __a =tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): __a =data_dir_format(_snake_case ) assert os.path.exists(_snake_case ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 __a =os.path.join(_snake_case , 'stage-1' ) __a ={ 'accelerator': accelerator, 'model_name_or_path': args.model_name_or_path, 'cache_dir': args.cache_dir, 'do_train': True, 'train_file': data_files['train'] if iteration == 0 else data_files['train_pseudo'], 'do_eval': True if args.eval_file is not None else False, 'eval_file': data_files['eval'], 'do_predict': True, 'infer_file': data_files['infer'], 'task_name': args.task_name, 'label_list': args.label_list, 'output_dir': current_output_dir, 'eval_metric': args.eval_metric, 'evaluation_strategy': args.evaluation_strategy, 'early_stopping_patience': args.early_stopping_patience, 'early_stopping_threshold': args.early_stopping_threshold, 'seed': args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(_snake_case , _snake_case ): arguments_dict.update({key: value} ) __a =os.path.join(_snake_case , 'best-checkpoint' , _snake_case ) if os.path.exists(_snake_case ): logger.info( 'Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.' , _snake_case , _snake_case , ) else: logger.info('***** Running self-training: iteration: %d, stage: 1 *****' , _snake_case ) finetune(**_snake_case ) accelerator.wait_for_everyone() assert os.path.exists(_snake_case ) logger.info('Self-training job completed: iteration: %d, stage: 1.' , _snake_case ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data __a =os.path.join(_snake_case , 'best-checkpoint' ) __a =os.path.join(_snake_case , 'stage-2' ) # Update arguments_dict __a =model_path __a =data_files['train'] __a =current_output_dir __a =os.path.join(_snake_case , 'best-checkpoint' , _snake_case ) if os.path.exists(_snake_case ): logger.info( 'Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.' , _snake_case , _snake_case , ) else: logger.info('***** Running self-training: iteration: %d, stage: 2 *****' , _snake_case ) finetune(**_snake_case ) accelerator.wait_for_everyone() assert os.path.exists(_snake_case ) logger.info('Self-training job completed: iteration: %d, stage: 2.' , _snake_case ) __a =iteration __a =data_dir_format(iteration + 1 ) __a =AutoConfig.from_pretrained(os.path.join(_snake_case , 'best-checkpoint' ) ) __a =config.idalabel __a =os.path.join(_snake_case , 'eval_results_best-checkpoint.json' ) __a =os.path.join(_snake_case , 'test_results_best-checkpoint.json' ) assert os.path.exists(_snake_case ) with open(_snake_case , 'r' ) as f: __a =float(json.load(_snake_case )[args.eval_metric] ) __a =os.path.join(_snake_case , 'infer_output_best-checkpoint.csv' ) assert os.path.exists(_snake_case ) # Loading the dataset from local csv or json files. __a =load_dataset(args.data_file_extension , data_files={'data': data_files['infer']} )['data'] __a =load_dataset('csv' , data_files={'data': infer_output_file} )['data'] if accelerator.is_main_process: os.makedirs(_snake_case , exist_ok=_snake_case ) shutil.copy(_snake_case , os.path.join(_snake_case , F'eval_results_iter-{iteration}.json' ) ) if os.path.exists(_snake_case ): shutil.copy(_snake_case , os.path.join(_snake_case , F'test_results_iter-{iteration}.json' ) ) create_pseudo_labeled_data(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) accelerator.wait_for_everyone() __a =os.path.join(_snake_case , F'train_pseudo.{args.data_file_extension}' ) if args.evaluation_strategy != IntervalStrategy.NO.value: __a =eval_result if best_iteration is None: __a =new_iteration __a =new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: __a =new_iteration __a =new_eval_result __a =0 else: if new_eval_result == best_eval_result: __a =new_iteration __a =new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: __a =True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info('Best iteration: %d' , _snake_case ) logger.info('Best evaluation result: %s = %f' , args.eval_metric , _snake_case ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_snake_case , F'eval_results_iter-{iteration}.json' ) , os.path.join(_snake_case , 'eval_results_best-iteration.json' ) , ) else: # Assume that the last iteration is the best logger.info('Best iteration: %d' , args.max_selftrain_iterations - 1 ) logger.info('Best evaluation result: %s = %f' , args.eval_metric , _snake_case ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_snake_case , F'eval_results_iter-{args.max_selftrain_iterations - 1}.json' ) , os.path.join(_snake_case , 'eval_results_best-iteration.json' ) , )
308
1
"""simple docstring""" from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__(self , SCREAMING_SNAKE_CASE__ ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = data SCREAMING_SNAKE_CASE__ : Node | None = None SCREAMING_SNAKE_CASE__ : Node | None = None def lowercase_ ( _snake_case ): # In Order traversal of the tree if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowercase_ ( _snake_case ): return 1 + max(depth_of_tree(tree.left ) ,depth_of_tree(tree.right ) ) if tree else 0 def lowercase_ ( _snake_case ): if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowercase_ ( ): # Main function for testing. SCREAMING_SNAKE_CASE__ : str = Node(1 ) SCREAMING_SNAKE_CASE__ : Dict = Node(2 ) SCREAMING_SNAKE_CASE__ : str = Node(3 ) SCREAMING_SNAKE_CASE__ : Any = Node(4 ) SCREAMING_SNAKE_CASE__ : List[Any] = Node(5 ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = Node(6 ) SCREAMING_SNAKE_CASE__ : Tuple = Node(7 ) SCREAMING_SNAKE_CASE__ : int = Node(8 ) SCREAMING_SNAKE_CASE__ : Any = Node(9 ) print(is_full_binary_tree(__a ) ) print(depth_of_tree(__a ) ) print("""Tree is: """ ) display(__a ) if __name__ == "__main__": main()
25
def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): snake_case_ : int = f"""Input value of [number={number}] must be an integer""" raise TypeError(__a ) if number < 0: return False snake_case_ : Dict = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
327
0
"""simple docstring""" import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) def snake_case ( A__ ): UpperCAmelCase_ : int = MobileNetVaConfig(layer_norm_eps=0.001 ) if "_quant" in model_name: raise ValueError("Quantized models are not supported." ) UpperCAmelCase_ : str = re.match(r"^mobilenet_v1_([^_]*)_([^_]*)$" ,A__ ) if matches: UpperCAmelCase_ : Tuple = float(matches[1] ) UpperCAmelCase_ : Tuple = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". UpperCAmelCase_ : Any = 10_01 UpperCAmelCase_ : Union[str, Any] = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[str] = "huggingface/label-files" UpperCAmelCase_ : Dict = json.load(open(hf_hub_download(A__ ,A__ ,repo_type="dataset" ) ,"r" ) ) UpperCAmelCase_ : Optional[int] = {int(A__ ) + 1: v for k, v in idalabel.items()} UpperCAmelCase_ : Optional[Any] = "background" UpperCAmelCase_ : Dict = idalabel UpperCAmelCase_ : str = {v: k for k, v in idalabel.items()} return config def snake_case ( ): UpperCAmelCase_ : str = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : Union[str, Any] = Image.open(requests.get(A__ ,stream=A__ ).raw ) return im @torch.no_grad() def snake_case ( A__ ,A__ ,A__ ,A__=False ): UpperCAmelCase_ : Dict = get_mobilenet_va_config(A__ ) # Load 🤗 model UpperCAmelCase_ : int = MobileNetVaForImageClassification(A__ ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(A__ ,A__ ,A__ ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor UpperCAmelCase_ : Optional[Any] = MobileNetVaImageProcessor( crop_size={"width": config.image_size, "height": config.image_size} ,size={"shortest_edge": config.image_size + 32} ,) UpperCAmelCase_ : List[str] = image_processor(images=prepare_img() ,return_tensors="pt" ) UpperCAmelCase_ : Any = model(**A__ ) UpperCAmelCase_ : List[Any] = outputs.logits assert logits.shape == (1, 10_01) if model_name == "mobilenet_v1_1.0_224": UpperCAmelCase_ : str = torch.tensor([-4.1739, -1.1233, 3.1205] ) elif model_name == "mobilenet_v1_0.75_192": UpperCAmelCase_ : Optional[int] = torch.tensor([-3.9440, -2.3141, -0.3333] ) else: UpperCAmelCase_ : Any = None if expected_logits is not None: assert torch.allclose(logits[0, :3] ,A__ ,atol=1e-4 ) Path(A__ ).mkdir(exist_ok=A__ ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(A__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(A__ ) if push_to_hub: print("Pushing to the hub..." ) UpperCAmelCase_ : str = "google/" + model_name image_processor.push_to_hub(A__ ) model.push_to_hub(A__ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''mobilenet_v1_1.0_224''', type=str, help='''Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.''', ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original TensorFlow checkpoint (.ckpt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCamelCase_ = parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
253
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class UpperCamelCase_ : def __init__( self : Optional[int] , lowerCAmelCase_ : int , lowerCAmelCase_ : MutableSequence[float] ) -> None: if len(lowerCAmelCase_ ) != degree + 1: raise ValueError( "The number of coefficients should be equal to the degree + 1." ) UpperCAmelCase_ : list[float] = list(lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = degree def __add__( self : int , lowerCAmelCase_ : Polynomial ) -> Polynomial: if self.degree > polynomial_a.degree: UpperCAmelCase_ : List[str] = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCAmelCase_ ) else: UpperCAmelCase_ : Optional[Any] = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCAmelCase_ ) def __sub__( self : Union[str, Any] , lowerCAmelCase_ : Polynomial ) -> Polynomial: return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : List[str] ) -> Polynomial: return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[Any] , lowerCAmelCase_ : Polynomial ) -> Polynomial: UpperCAmelCase_ : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase_ : int | float ) -> int | float: UpperCAmelCase_ : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Tuple ) -> str: UpperCAmelCase_ : str = "" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCAmelCase_ ) return polynomial def __repr__( self : Union[str, Any] ) -> str: return self.__str__() def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Polynomial: UpperCAmelCase_ : list[float] = [0] * self.degree for i in range(self.degree ): UpperCAmelCase_ : List[Any] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : int | float = 0 ) -> Polynomial: UpperCAmelCase_ : list[float] = [0] * (self.degree + 2) UpperCAmelCase_ : List[Any] = constant for i in range(self.degree + 1 ): UpperCAmelCase_ : Union[str, Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCAmelCase_ ) def __eq__( self : Union[str, Any] , lowerCAmelCase_ : object ) -> bool: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Tuple , lowerCAmelCase_ : object ) -> bool: return not self.__eq__(lowerCAmelCase_ )
253
1
"""simple docstring""" import numpy as np def lowerCamelCase__ ( _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Tuple ) -> List[str]: lowerCamelCase_ = int(np.ceil((x_end - xa) / h ) ) lowerCamelCase_ = np.zeros((n + 1,) ) lowerCamelCase_ = ya lowerCamelCase_ = xa for k in range(_lowerCamelCase ): lowerCamelCase_ = f(_lowerCamelCase , y[k] ) lowerCamelCase_ = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) lowerCamelCase_ = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) lowerCamelCase_ = f(x + h , y[k] + h * ka ) lowerCamelCase_ = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
183
"""simple docstring""" from cva import destroyAllWindows, imread, imshow, waitKey def lowerCamelCase__ ( _lowerCamelCase : Tuple ) -> Dict: # getting number of pixels in the image lowerCamelCase_ , lowerCamelCase_ = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(_lowerCamelCase ): for j in range(_lowerCamelCase ): lowerCamelCase_ = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image _SCREAMING_SNAKE_CASE : List[Any] = imread('''image_data/lena.jpg''', 1) # convert to its negative _SCREAMING_SNAKE_CASE : List[Any] = convert_to_negative(img) # show result image imshow('''negative of original image''', img) waitKey(0) destroyAllWindows()
183
1
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder lowercase__ : List[Any] = '''base_with_context''' def __lowercase ( _a , _a ): snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights['''token_embedder''']['''embedding'''] ) ) snake_case_ : Optional[int] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=snake_case_ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case_ : Any = weights[f"layers_{lyr_num}"] snake_case_ : Optional[Any] = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) snake_case_ : str = ly_weight["""attention"""] snake_case_ : Dict = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) snake_case_ : str = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) snake_case_ : Any = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) snake_case_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) snake_case_ : Any = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) snake_case_ : Dict = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def __lowercase ( _a , _a ): snake_case_ : Dict = nn.Parameter(torch.FloatTensor(weights['''input_proj''']['''kernel'''].T ) ) snake_case_ : int = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=snake_case_ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case_ : Any = weights[f"layers_{lyr_num}"] snake_case_ : Optional[Any] = ly_weight["""attention"""] snake_case_ : Any = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) snake_case_ : str = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) snake_case_ : Tuple = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) snake_case_ : str = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) snake_case_ : str = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) snake_case_ : Optional[Any] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def __lowercase ( _a , _a ): snake_case_ : Optional[Any] = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense0''']['''kernel'''].T ) ) snake_case_ : List[str] = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense1''']['''kernel'''].T ) ) snake_case_ : List[str] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=snake_case_ ) snake_case_ : List[Any] = nn.Parameter( torch.FloatTensor(weights['''continuous_inputs_projection''']['''kernel'''].T ) ) for lyr_num, lyr in enumerate(model.decoders ): snake_case_ : Union[str, Any] = weights[f"layers_{lyr_num}"] snake_case_ : int = nn.Parameter( torch.FloatTensor(ly_weight['''pre_self_attention_layer_norm''']['''scale'''] ) ) snake_case_ : int = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_0''']['''DenseGeneral_0''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = ly_weight["""self_attention"""] snake_case_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) snake_case_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) snake_case_ : List[str] = ly_weight["""MultiHeadDotProductAttention_0"""] snake_case_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) snake_case_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) snake_case_ : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight['''pre_cross_attention_layer_norm''']['''scale'''] ) ) snake_case_ : Any = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) snake_case_ : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_1''']['''DenseGeneral_0''']['''kernel'''].T ) ) snake_case_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) snake_case_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) snake_case_ : str = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) snake_case_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights['''decoder_norm''']['''scale'''] ) ) snake_case_ : Any = nn.Parameter(torch.FloatTensor(weights['''spec_out_dense''']['''kernel'''].T ) ) return model def __lowercase ( _a ): snake_case_ : str = checkpoints.load_tax_checkpoint(args.checkpoint_path ) snake_case_ : Optional[Any] = jnp.tree_util.tree_map(onp.array , snake_case_ ) snake_case_ : Optional[Any] = [ """from __gin__ import dynamic_registration""", """from music_spectrogram_diffusion.models.diffusion import diffusion_utils""", """diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0""", """diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()""", ] snake_case_ : Optional[Any] = os.path.join(args.checkpoint_path , '''..''' , '''config.gin''' ) snake_case_ : Dict = inference.parse_training_gin_file(snake_case_ , snake_case_ ) snake_case_ : Optional[Any] = inference.InferenceModel(args.checkpoint_path , snake_case_ ) snake_case_ : List[Any] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' , variance_type='''fixed_large''' ) snake_case_ : int = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['''inputs'''] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) snake_case_ : str = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['''targets_context'''] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) snake_case_ : Dict = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['''targets_context'''] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) snake_case_ : str = load_notes_encoder(ta_checkpoint['''target''']['''token_encoder'''] , snake_case_ ) snake_case_ : Optional[Any] = load_continuous_encoder(ta_checkpoint['''target''']['''continuous_encoder'''] , snake_case_ ) snake_case_ : Union[str, Any] = load_decoder(ta_checkpoint['''target''']['''decoder'''] , snake_case_ ) snake_case_ : List[str] = OnnxRuntimeModel.from_pretrained('''kashif/soundstream_mel_decoder''' ) snake_case_ : List[Any] = SpectrogramDiffusionPipeline( notes_encoder=snake_case_ , continuous_encoder=snake_case_ , decoder=snake_case_ , scheduler=snake_case_ , melgan=snake_case_ , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": lowercase__ : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--output_path''', default=None, type=str, required=True, help='''Path to the converted model.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument( '''--checkpoint_path''', default=f'{MODEL}/checkpoint_500000', type=str, required=False, help='''Path to the original jax model checkpoint.''', ) lowercase__ : Optional[Any] = parser.parse_args() main(args)
355
"""simple docstring""" from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class _UpperCAmelCase ( lowerCAmelCase__): def __init__( self : List[Any] , lowercase_ : NestedDataStructureLike[PathLike] , lowercase_ : Optional[NamedSplit] = None , lowercase_ : Optional[Features] = None , lowercase_ : str = None , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : Optional[int] = None , **lowercase_ : Optional[Any] , ): super().__init__( lowercase_ , split=lowercase_ , features=lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ , streaming=lowercase_ , num_proc=lowercase_ , **lowercase_ , ) snake_case_ : List[Any] = path_or_paths if isinstance(lowercase_ , lowercase_ ) else {self.split: path_or_paths} snake_case_ : str = Text( cache_dir=lowercase_ , data_files=lowercase_ , features=lowercase_ , **lowercase_ , ) def _snake_case ( self : Any ): # Build iterable dataset if self.streaming: snake_case_ : Union[str, Any] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: snake_case_ : List[Any] = None snake_case_ : Optional[Any] = None snake_case_ : str = None snake_case_ : Optional[int] = None self.builder.download_and_prepare( download_config=lowercase_ , download_mode=lowercase_ , verification_mode=lowercase_ , base_path=lowercase_ , num_proc=self.num_proc , ) snake_case_ : Union[str, Any] = self.builder.as_dataset( split=self.split , verification_mode=lowercase_ , in_memory=self.keep_in_memory ) return dataset
155
0
def a_ ( __lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[str]=False ) -> str: if isinstance(__a , __a ) and isinstance(__a , __a ): _snake_case = len(set_a.intersection(__a ) ) if alternative_union: _snake_case = len(__a ) + len(__a ) else: _snake_case = len(set_a.union(__a ) ) return intersection / union if isinstance(__a , (list, tuple) ) and isinstance(__a , (list, tuple) ): _snake_case = [element for element in set_a if element in set_b] if alternative_union: _snake_case = len(__a ) + len(__a ) return len(__a ) / union else: _snake_case = set_a + [element for element in set_b if element not in set_a] return len(__a ) / len(__a ) return len(__a ) / len(__a ) return None if __name__ == "__main__": _lowerCamelCase : Union[str, Any] = {'''a''', '''b''', '''c''', '''d''', '''e'''} _lowerCamelCase : str = {'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
282
'''simple docstring''' def a ( __a , __a ) -> int: '''simple docstring''' if len(__a ) != len(__a ): raise ValueError('''String lengths must match!''' ) UpperCamelCase__ :Union[str, Any] = 0 for chara, chara in zip(__a , __a ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
97
0
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class lowercase_ ( unittest.TestCase ): '''simple docstring''' __snake_case = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __lowerCAmelCase ( self : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int] ) ->Union[str, Any]: """simple docstring""" a = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) a = VideoClassificationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase , top_k=2 ) a = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __lowerCAmelCase ( self : int , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict ) ->int: """simple docstring""" for example in examples: a = video_classifier(__UpperCAmelCase ) self.assertEqual( __UpperCAmelCase , [ {'''score''': ANY(__UpperCAmelCase ), '''label''': ANY(__UpperCAmelCase )}, {'''score''': ANY(__UpperCAmelCase ), '''label''': ANY(__UpperCAmelCase )}, ] , ) @require_torch def __lowerCAmelCase ( self : str ) ->Dict: """simple docstring""" a = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' a = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) a = pipeline( '''video-classification''' , model=__UpperCAmelCase , feature_extractor=__UpperCAmelCase , frame_sampling_rate=4 ) a = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) a = video_classifier(__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) a = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __lowerCAmelCase ( self : Any ) ->str: """simple docstring""" pass
370
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class lowercase_ ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : Optional[int] ) ->Tuple: """simple docstring""" a = tempfile.mkdtemp() a = BlipImageProcessor() a = GPTaTokenizer.from_pretrained('''hf-internal-testing/tiny-random-GPT2Model''' ) a = BertTokenizerFast.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) a = InstructBlipProcessor(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Optional[Any] , **__UpperCAmelCase : Tuple ) ->List[str]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).tokenizer def __lowerCAmelCase ( self : int , **__UpperCAmelCase : str ) ->List[str]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).image_processor def __lowerCAmelCase ( self : Optional[Any] , **__UpperCAmelCase : Any ) ->Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).qformer_tokenizer def __lowerCAmelCase ( self : str ) ->Tuple: """simple docstring""" shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : Optional[int] ) ->str: """simple docstring""" a = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] a = [Image.fromarray(np.moveaxis(__UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __lowerCAmelCase ( self : Optional[Any] ) ->List[str]: """simple docstring""" a = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) a = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) a = self.get_image_processor(do_normalize=__UpperCAmelCase , padding_value=1.0 ) a = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __UpperCAmelCase ) self.assertIsInstance(processor.qformer_tokenizer , __UpperCAmelCase ) def __lowerCAmelCase ( self : Optional[Any] ) ->Any: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = self.get_qformer_tokenizer() a = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) a = self.prepare_image_inputs() a = image_processor(__UpperCAmelCase , return_tensors='''np''' ) a = processor(images=__UpperCAmelCase , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __lowerCAmelCase ( self : List[str] ) ->List[Any]: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = self.get_qformer_tokenizer() a = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) a = '''lower newer''' a = processor(text=__UpperCAmelCase ) a = tokenizer(__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase ) a = qformer_tokenizer(__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['''qformer_''' + key] ) def __lowerCAmelCase ( self : Dict ) ->Optional[Any]: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = self.get_qformer_tokenizer() a = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) a = '''lower newer''' a = self.prepare_image_inputs() a = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual( list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''qformer_input_ids''', '''qformer_attention_mask''', '''pixel_values'''] , ) # test if it raises when no input is passed with pytest.raises(__UpperCAmelCase ): processor() def __lowerCAmelCase ( self : Dict ) ->List[Any]: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = self.get_qformer_tokenizer() a = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) a = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] a = processor.batch_decode(__UpperCAmelCase ) a = tokenizer.batch_decode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def __lowerCAmelCase ( self : Union[str, Any] ) ->str: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = self.get_qformer_tokenizer() a = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) a = '''lower newer''' a = self.prepare_image_inputs() a = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual( list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''qformer_input_ids''', '''qformer_attention_mask''', '''pixel_values'''] , )
26
0
from collections import defaultdict class __magic_name__ : """simple docstring""" def __init__( self :Tuple , snake_case :List[Any] , snake_case :Union[str, Any] ): '''simple docstring''' A_ : Optional[int] = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 A_ : str = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(snake_case ) ) ] A_ : Dict = defaultdict(snake_case ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 A_ : List[str] = (1 << len(snake_case )) - 1 def SCREAMING_SNAKE_CASE ( self :Tuple , snake_case :Tuple , snake_case :List[str] ): '''simple docstring''' if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement A_ : Any = self.count_ways_until(snake_case , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. A_ : int = total_ways_util return self.dp[mask][task_no] def SCREAMING_SNAKE_CASE ( self :Any , snake_case :Optional[Any] ): '''simple docstring''' for i in range(len(snake_case ) ): for j in task_performed[i]: self.task[j].append(snake_case ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": _lowerCAmelCase : int = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. _lowerCAmelCase : int = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
300
import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = (DDPMScheduler,) def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , **snake_case :str ): '''simple docstring''' A_ : Dict = { "num_train_timesteps": 1_000, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", "variance_type": "fixed_small", "clip_sample": True, } config.update(**snake_case ) return config def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case , beta_end=snake_case ) def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' self.check_over_configs(thresholding=snake_case ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case , prediction_type=snake_case , sample_max_value=snake_case , ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' A_ : Tuple = self.scheduler_classes[0] A_ : List[str] = self.get_scheduler_config() A_ : List[str] = scheduler_class(**snake_case ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.00979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1e-5 def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : int = self.scheduler_classes[0] A_ : List[str] = self.get_scheduler_config() A_ : int = scheduler_class(**snake_case ) A_ : Tuple = len(snake_case ) A_ : List[str] = self.dummy_model() A_ : Optional[Any] = self.dummy_sample_deter A_ : List[str] = torch.manual_seed(0 ) for t in reversed(range(snake_case ) ): # 1. predict noise residual A_ : Tuple = model(snake_case , snake_case ) # 2. predict previous mean of sample x_t-1 A_ : Dict = scheduler.step(snake_case , snake_case , snake_case , generator=snake_case ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance A_ : Optional[int] = pred_prev_sample A_ : Tuple = torch.sum(torch.abs(snake_case ) ) A_ : str = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 258.9606 ) < 1e-2 assert abs(result_mean.item() - 0.3372 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : Optional[int] = self.scheduler_classes[0] A_ : int = self.get_scheduler_config(prediction_type="v_prediction" ) A_ : List[str] = scheduler_class(**snake_case ) A_ : int = len(snake_case ) A_ : Dict = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Any = torch.manual_seed(0 ) for t in reversed(range(snake_case ) ): # 1. predict noise residual A_ : Optional[int] = model(snake_case , snake_case ) # 2. predict previous mean of sample x_t-1 A_ : Tuple = scheduler.step(snake_case , snake_case , snake_case , generator=snake_case ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance A_ : List[str] = pred_prev_sample A_ : Optional[Any] = torch.sum(torch.abs(snake_case ) ) A_ : List[str] = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 202.0296 ) < 1e-2 assert abs(result_mean.item() - 0.2631 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : str = self.scheduler_classes[0] A_ : Optional[Any] = self.get_scheduler_config() A_ : Dict = scheduler_class(**snake_case ) A_ : Optional[int] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case ) A_ : Optional[int] = scheduler.timesteps for i, timestep in enumerate(snake_case ): if i == len(snake_case ) - 1: A_ : str = -1 else: A_ : List[str] = timesteps[i + 1] A_ : Optional[int] = scheduler.previous_timestep(snake_case ) A_ : List[str] = prev_t.item() self.assertEqual(snake_case , snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : Optional[Any] = self.scheduler_classes[0] A_ : int = self.get_scheduler_config() A_ : Tuple = scheduler_class(**snake_case ) A_ : List[str] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case , msg="`custom_timesteps` must be in descending order." ): scheduler.set_timesteps(timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Any = self.scheduler_classes[0] A_ : Union[str, Any] = self.get_scheduler_config() A_ : Optional[int] = scheduler_class(**snake_case ) A_ : Union[str, Any] = [100, 87, 50, 1, 0] A_ : Optional[int] = len(snake_case ) with self.assertRaises(snake_case , msg="Can only pass one of `num_inference_steps` or `custom_timesteps`." ): scheduler.set_timesteps(num_inference_steps=snake_case , timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Optional[Any] = self.get_scheduler_config() A_ : Optional[int] = scheduler_class(**snake_case ) A_ : Optional[int] = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case , msg="`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}" , ): scheduler.set_timesteps(timesteps=snake_case )
300
1
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _A = '''tiny-wmt19-en-ru''' # Build # borrowed from a test _A = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _A = dict(zip(vocab, range(len(vocab)))) _A = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: _A = Path(tmpdirname) _A = build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] _A = build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] _A = build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) _A = FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _A = FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _A = FSMTForConditionalGeneration(config) print(F'''num of params {tiny_model.num_parameters()}''') # Test _A = tokenizer(['''Making tiny model'''], return_tensors='''pt''') _A = tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'''Generated {mname_tiny}''') # Upload # transformers-cli upload tiny-wmt19-en-ru
365
from __future__ import annotations def lowerCamelCase__ ( a__ : list[list[int]] ) -> int: # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(a__ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(a__ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
261
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCAmelCase : List[Any] = { 'configuration_gpt_neox_japanese': ['GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoXJapaneseConfig'], 'tokenization_gpt_neox_japanese': ['GPTNeoXJapaneseTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Tuple = [ 'GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTNeoXJapaneseForCausalLM', 'GPTNeoXJapaneseLayer', 'GPTNeoXJapaneseModel', 'GPTNeoXJapanesePreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCAmelCase : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
88
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __magic_name__ ( __a : Optional[int] , __a : Union[str, Any] , __a : Union[str, Any]=1_024 , __a : str=1_024 , __a : Optional[Any]=False , **__a : Tuple ): '''simple docstring''' UpperCamelCase__ = AutoTokenizer.from_pretrained(__a ) UpperCamelCase__ = SeqaSeqDataset(__a , __a , __a , __a , type_path="""train""" , **__a ) UpperCamelCase__ = tok.pad_token_id def get_lens(__a : Optional[int] ): UpperCamelCase__ = tqdm( DataLoader(__a , batch_size=512 , num_workers=8 , shuffle=__a , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) UpperCamelCase__ = [] for batch in dl: UpperCamelCase__ = batch["""input_ids"""].ne(__a ).sum(1 ).tolist() UpperCamelCase__ = batch["""labels"""].ne(__a ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__a , __a ): max_lens.append(max(__a , __a ) ) else: max_lens.extend(__a ) return max_lens UpperCamelCase__ = get_lens(__a ) UpperCamelCase__ = SeqaSeqDataset(__a , __a , __a , __a , type_path="""val""" , **__a ) UpperCamelCase__ = get_lens(__a ) pickle_save(__a , train_ds.len_file ) pickle_save(__a , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
244
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __a (UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :Optional[Any] = CycleDiffusionPipeline _SCREAMING_SNAKE_CASE :List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """negative_prompt""", """height""", """width""", """negative_prompt_embeds""", } _SCREAMING_SNAKE_CASE :Any = PipelineTesterMixin.required_optional_params - {"""latents"""} _SCREAMING_SNAKE_CASE :Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""}) _SCREAMING_SNAKE_CASE :Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS _SCREAMING_SNAKE_CASE :Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def _a ( self ) -> Any: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) SCREAMING_SNAKE_CASE__ : Tuple = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , num_train_timesteps=1_000 , clip_sample=_a , set_alpha_to_one=_a , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) SCREAMING_SNAKE_CASE__ : int = CLIPTextModel(_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) SCREAMING_SNAKE_CASE__ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def _a ( self , _a , _a=0 ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) SCREAMING_SNAKE_CASE__ : Dict = image / 2 + 0.5 if str(_a ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ : int = torch.manual_seed(_a ) else: SCREAMING_SNAKE_CASE__ : List[Any] = torch.Generator(device=_a ).manual_seed(_a ) SCREAMING_SNAKE_CASE__ : str = { """prompt""": """An astronaut riding an elephant""", """source_prompt""": """An astronaut riding a horse""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """eta""": 0.1, """strength""": 0.8, """guidance_scale""": 3, """source_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ : List[str] = self.get_dummy_components() SCREAMING_SNAKE_CASE__ : int = CycleDiffusionPipeline(**_a ) SCREAMING_SNAKE_CASE__ : str = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : List[Any] = self.get_dummy_inputs(_a ) SCREAMING_SNAKE_CASE__ : int = pipe(**_a ) SCREAMING_SNAKE_CASE__ : int = output.images SCREAMING_SNAKE_CASE__ : Dict = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) SCREAMING_SNAKE_CASE__ : Tuple = np.array([0.4_459, 0.4_943, 0.4_544, 0.6_643, 0.5_474, 0.4_327, 0.5_701, 0.5_959, 0.5_179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_dummy_components() for name, module in components.items(): if hasattr(_a , """half""" ): SCREAMING_SNAKE_CASE__ : int = module.half() SCREAMING_SNAKE_CASE__ : Tuple = CycleDiffusionPipeline(**_a ) SCREAMING_SNAKE_CASE__ : Any = pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) SCREAMING_SNAKE_CASE__ : Dict = self.get_dummy_inputs(_a ) SCREAMING_SNAKE_CASE__ : Any = pipe(**_a ) SCREAMING_SNAKE_CASE__ : Tuple = output.images SCREAMING_SNAKE_CASE__ : Any = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) SCREAMING_SNAKE_CASE__ : List[Any] = np.array([0.3_506, 0.4_543, 0.446, 0.4_575, 0.5_195, 0.4_155, 0.5_273, 0.518, 0.4_116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def _a ( self ) -> int: """simple docstring""" return super().test_save_load_local() @unittest.skip("""non-deterministic pipeline""" ) def _a ( self ) -> Any: """simple docstring""" return super().test_inference_batch_single_identical() @skip_mps def _a ( self ) -> Optional[Any]: """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def _a ( self ) -> Union[str, Any]: """simple docstring""" return super().test_save_load_optional_components() @skip_mps def _a ( self ) -> List[str]: """simple docstring""" return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class __a (unittest.TestCase): '''simple docstring''' def _a ( self ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) SCREAMING_SNAKE_CASE__ : Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy""" ) SCREAMING_SNAKE_CASE__ : Any = init_image.resize((512, 512) ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = """CompVis/stable-diffusion-v1-4""" SCREAMING_SNAKE_CASE__ : List[Any] = DDIMScheduler.from_pretrained(_a , subfolder="""scheduler""" ) SCREAMING_SNAKE_CASE__ : str = CycleDiffusionPipeline.from_pretrained( _a , scheduler=_a , safety_checker=_a , torch_dtype=torch.floataa , revision="""fp16""" ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) pipe.enable_attention_slicing() SCREAMING_SNAKE_CASE__ : Tuple = """A black colored car""" SCREAMING_SNAKE_CASE__ : List[Any] = """A blue colored car""" SCREAMING_SNAKE_CASE__ : Any = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : str = pipe( prompt=_a , source_prompt=_a , image=_a , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_a , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : Any = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) SCREAMING_SNAKE_CASE__ : List[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy""" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = init_image.resize((512, 512) ) SCREAMING_SNAKE_CASE__ : Tuple = """CompVis/stable-diffusion-v1-4""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = DDIMScheduler.from_pretrained(_a , subfolder="""scheduler""" ) SCREAMING_SNAKE_CASE__ : int = CycleDiffusionPipeline.from_pretrained(_a , scheduler=_a , safety_checker=_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) pipe.enable_attention_slicing() SCREAMING_SNAKE_CASE__ : int = """A black colored car""" SCREAMING_SNAKE_CASE__ : Optional[Any] = """A blue colored car""" SCREAMING_SNAKE_CASE__ : Tuple = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Dict = pipe( prompt=_a , source_prompt=_a , image=_a , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_a , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = output.images assert np.abs(image - expected_image ).max() < 2E-2
56
"""simple docstring""" import math from collections.abc import Iterator from itertools import takewhile def _lowercase ( __lowerCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _lowercase ( ) -> Iterator[int]: SCREAMING_SNAKE_CASE__ : List[Any] = 2 while True: if is_prime(__lowerCAmelCase ): yield num num += 1 def _lowercase ( __lowerCAmelCase = 200_0000 ) -> int: return sum(takewhile(lambda __lowerCAmelCase : x < n , prime_generator() ) ) if __name__ == "__main__": print(f'{solution() = }')
56
1