code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
"""simple docstring""" import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class lowercase: '''simple docstring''' def __init__( self: Any, a_: Union[str, Any], a_: Dict=13, a_: Optional[Any]=32, a_: Any=2, a_: Any=3, a_: Optional[Any]=16, a_: List[str]=[1, 2, 1], a_: int=[2, 2, 4], a_: Dict=2, a_: Optional[int]=2.0, a_: Union[str, Any]=True, a_: Optional[Any]=0.0, a_: Optional[int]=0.0, a_: Union[str, Any]=0.1, a_: str="gelu", a_: int=False, a_: Union[str, Any]=True, a_: Dict=0.02, a_: List[Any]=1E-5, a_: int=True, a_: Union[str, Any]=None, a_: Optional[int]=True, a_: List[Any]=10, a_: Tuple=8, a_: Optional[Any]=["stage1", "stage2", "stage3"], a_: Union[str, Any]=[1, 2, 3], ): '''simple docstring''' _snake_case : str = parent _snake_case : Optional[int] = batch_size _snake_case : Any = image_size _snake_case : int = patch_size _snake_case : Union[str, Any] = num_channels _snake_case : int = embed_dim _snake_case : Optional[Any] = depths _snake_case : Tuple = num_heads _snake_case : Union[str, Any] = window_size _snake_case : List[Any] = mlp_ratio _snake_case : Union[str, Any] = qkv_bias _snake_case : List[Any] = hidden_dropout_prob _snake_case : Dict = attention_probs_dropout_prob _snake_case : Union[str, Any] = drop_path_rate _snake_case : str = hidden_act _snake_case : Union[str, Any] = use_absolute_embeddings _snake_case : Optional[Any] = patch_norm _snake_case : Any = layer_norm_eps _snake_case : Union[str, Any] = initializer_range _snake_case : Union[str, Any] = is_training _snake_case : Optional[Any] = scope _snake_case : Union[str, Any] = use_labels _snake_case : Union[str, Any] = type_sequence_label_size _snake_case : str = encoder_stride _snake_case : List[Any] = out_features _snake_case : Tuple = out_indices def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case : Tuple = None if self.use_labels: _snake_case : Optional[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) _snake_case : List[str] = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self: str ): '''simple docstring''' return MaskFormerSwinConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, embed_dim=self.embed_dim, depths=self.depths, num_heads=self.num_heads, window_size=self.window_size, mlp_ratio=self.mlp_ratio, qkv_bias=self.qkv_bias, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, drop_path_rate=self.drop_path_rate, hidden_act=self.hidden_act, use_absolute_embeddings=self.use_absolute_embeddings, path_norm=self.patch_norm, layer_norm_eps=self.layer_norm_eps, initializer_range=self.initializer_range, encoder_stride=self.encoder_stride, out_features=self.out_features, out_indices=self.out_indices, ) def UpperCamelCase_ ( self: str, a_: List[str], a_: List[str], a_: str ): '''simple docstring''' _snake_case : str = MaskFormerSwinModel(config=a_ ) model.to(a_ ) model.eval() _snake_case : Any = model(a_ ) _snake_case : List[str] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _snake_case : List[str] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Tuple, a_: Any ): '''simple docstring''' _snake_case : int = MaskFormerSwinBackbone(config=a_ ) model.to(a_ ) model.eval() _snake_case : str = model(a_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ), len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ), [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ), len(config.out_features ) ) self.parent.assertListEqual(model.channels, [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(a_ ): _snake_case : Optional[Any] = ["""stem"""] _snake_case : Tuple = MaskFormerSwinBackbone(config=a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Any = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case : List[Any] = config_and_inputs _snake_case : Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowercase( __a , __a , unittest.TestCase ): '''simple docstring''' lowercase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : List[Any] = MaskFormerSwinModelTester(self ) _snake_case : List[str] = ConfigTester(self, config_class=a_, embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def UpperCamelCase_ ( self: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*a_ ) @unittest.skip("""Swin does not use inputs_embeds""" ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' pass @unittest.skip("""Swin does not support feedforward chunking""" ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case , _snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Optional[int] = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) _snake_case : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_, nn.Linear ) ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case , _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Tuple = model_class(a_ ) _snake_case : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : Union[str, Any] = [*signature.parameters.keys()] _snake_case : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1], a_ ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: str, a_: Dict, a_: Union[str, Any], a_: Union[str, Any] ): '''simple docstring''' _snake_case : Any = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): _snake_case : Optional[Any] = model(**self._prepare_for_class(a_, a_ ) ) _snake_case : Optional[Any] = outputs.hidden_states _snake_case : Any = getattr( self.model_tester, """expected_num_hidden_layers""", len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a_ ), a_ ) # Swin has a different seq_length _snake_case : int = ( config.patch_size if isinstance(config.patch_size, collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _snake_case : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ), [num_patches, self.model_tester.embed_dim], ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case , _snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Any = ( self.model_tester.image_size if isinstance(self.model_tester.image_size, collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _snake_case : Tuple = True self.check_hidden_states_output(a_, a_, a_, a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : Optional[Any] = True self.check_hidden_states_output(a_, a_, a_, a_ ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case , _snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : str = 3 _snake_case : List[str] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size, collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _snake_case : List[Any] = ( config.patch_size if isinstance(config.patch_size, collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _snake_case : List[str] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _snake_case : Tuple = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _snake_case : Any = True self.check_hidden_states_output(a_, a_, a_, (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : Dict = True self.check_hidden_states_output(a_, a_, a_, (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def UpperCamelCase_ ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def UpperCamelCase_ ( self: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case , _snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(a_: List[str] ): _snake_case : Union[str, Any] = 0 return t def check_equivalence(a_: List[Any], a_: List[Any], a_: List[str], a_: List[str]={} ): with torch.no_grad(): _snake_case : Any = model(**a_, return_dict=a_, **a_ ) _snake_case : int = model(**a_, return_dict=a_, **a_ ).to_tuple() def recursive_check(a_: Union[str, Any], a_: Tuple ): if isinstance(a_, (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(a_, a_ ): recursive_check(a_, a_ ) elif isinstance(a_, a_ ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values(), dict_object.values() ): recursive_check(a_, a_ ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(a_ ), set_nan_tensor_to_zero(a_ ), atol=1E-5 ), msg=( """Tuple and dict output are not equal. Difference:""" f" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" f" {torch.isnan(a_ ).any()} and `inf`: {torch.isinf(a_ )}. Dict has" f" `nan`: {torch.isnan(a_ ).any()} and `inf`: {torch.isinf(a_ )}." ), ) recursive_check(a_, a_ ) for model_class in self.all_model_classes: _snake_case : Tuple = model_class(a_ ) model.to(a_ ) model.eval() _snake_case : int = self._prepare_for_class(a_, a_ ) _snake_case : str = self._prepare_for_class(a_, a_ ) check_equivalence(a_, a_, a_ ) _snake_case : str = self._prepare_for_class(a_, a_, return_labels=a_ ) _snake_case : str = self._prepare_for_class(a_, a_, return_labels=a_ ) check_equivalence(a_, a_, a_ ) _snake_case : Tuple = self._prepare_for_class(a_, a_ ) _snake_case : str = self._prepare_for_class(a_, a_ ) check_equivalence(a_, a_, a_, {"""output_hidden_states""": True} ) _snake_case : int = self._prepare_for_class(a_, a_, return_labels=a_ ) _snake_case : Optional[int] = self._prepare_for_class(a_, a_, return_labels=a_ ) check_equivalence(a_, a_, a_, {"""output_hidden_states""": True} ) @require_torch class lowercase( unittest.TestCase , __a ): '''simple docstring''' lowercase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase__ = MaskFormerSwinConfig def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : List[Any] = MaskFormerSwinModelTester(self ) def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case , _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: _snake_case : Any = backbone_class(a_ ) backbone.to(a_ ) backbone.eval() _snake_case : Union[str, Any] = backbone(**a_ ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps, a_ ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps, backbone.channels ): self.assertTrue(feature_map.shape[:2], (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _snake_case : List[str] = backbone(**a_, output_hidden_states=a_ ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ), len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:], backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _snake_case , _snake_case , _snake_case : Any = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels), (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _snake_case : Dict = backbone(**a_, output_attentions=a_ ) self.assertIsNotNone(outputs.attentions )
64
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = CodeGenTokenizer lowercase__ = CodeGenTokenizerFast lowercase__ = True lowercase__ = {"add_prefix_space": True} lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : Tuple = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] _snake_case : Tuple = dict(zip(a_, range(len(a_ ) ) ) ) _snake_case : str = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _snake_case : List[Any] = {"""unk_token""": """<unk>"""} _snake_case : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(a_ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(a_ ) ) def UpperCamelCase_ ( self: Any, **a_: int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Any, **a_: str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Dict ): '''simple docstring''' _snake_case : Union[str, Any] = """lower newer""" _snake_case : Tuple = """lower newer""" return input_text, output_text def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = CodeGenTokenizer(self.vocab_file, self.merges_file, **self.special_tokens_map ) _snake_case : Optional[Any] = """lower newer""" _snake_case : Optional[int] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _snake_case : int = tokenizer.tokenize(a_, add_prefix_space=a_ ) self.assertListEqual(a_, a_ ) _snake_case : str = tokens + [tokenizer.unk_token] _snake_case : Optional[int] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : int = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : Dict = """lower newer""" # Testing tokenization _snake_case : Dict = tokenizer.tokenize(a_, add_prefix_space=a_ ) _snake_case : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids without special tokens _snake_case : Optional[Any] = tokenizer.encode(a_, add_special_tokens=a_, add_prefix_space=a_ ) _snake_case : Tuple = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids with special tokens _snake_case : Tuple = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : int = tokenizer.encode(a_, add_prefix_space=a_ ) _snake_case : Optional[Any] = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) # Testing the unknown token _snake_case : Tuple = tokens + [rust_tokenizer.unk_token] _snake_case : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: List[Any]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) # Simple input _snake_case : Any = """This is a simple input""" _snake_case : Optional[int] = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") _snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) # Pair input self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname, pad_token="""<pad>""" ) # Simple input _snake_case : List[Any] = """This is a simple input""" _snake_case : int = ["""This is a simple input looooooooong""", """This is a simple input"""] _snake_case : Any = ("""This is a simple input""", """This is a pair""") _snake_case : str = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] _snake_case : str = tokenizer.pad_token_id _snake_case : Optional[int] = tokenizer(a_, padding="""max_length""", max_length=30, return_tensors="""np""" ) _snake_case : Dict = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) _snake_case : Tuple = tokenizer(*a_, padding="""max_length""", max_length=60, return_tensors="""np""" ) _snake_case : Optional[Any] = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1], 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1], 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1], 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1], 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = """$$$""" _snake_case : List[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname, bos_token=a_, add_bos_token=a_ ) _snake_case : str = """This is a simple input""" _snake_case : int = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Tuple = tokenizer(a_ ) _snake_case : Optional[Any] = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0], a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _snake_case : Optional[int] = tokenizer.decode(out_s.input_ids ) _snake_case : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0], a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) _snake_case : Dict = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" _snake_case : Union[str, Any] = """\nif len_a > len_b: result = a\nelse: result = b""" _snake_case : Optional[Any] = tokenizer.encode(a_ ) _snake_case : Dict = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] _snake_case : Optional[Any] = tokenizer.decode(a_, truncate_before_pattern=a_ ) self.assertEqual(a_, a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass
64
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def UpperCAmelCase__ (snake_case__ : List[Any]=None ): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(add_help=snake_case__ , allow_abbrev=snake_case__ ) # The main config parser _snake_case : Any = config_command_parser(snake_case__ ) # The subparser to add commands to _snake_case : Optional[Any] = config_parser.add_subparsers(title="""subcommands""" , dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(snake_case__ , parents=[parent_parser] ) update_command_parser(snake_case__ , parents=[parent_parser] ) return config_parser def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = get_config_parser() _snake_case : Optional[int] = config_parser.parse_args() if not hasattr(snake_case__ , """func""" ): config_parser.print_help() exit(1 ) # Run args.func(snake_case__ ) if __name__ == "__main__": main()
64
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ = re.compile(r'''\s+''') def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" return {"hash": hashlib.mda(re.sub(snake_case__ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : Any = [len(snake_case__ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(snake_case__ ), "line_max": max(snake_case__ )} def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : List[Any] ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : List[str]=5 ): """simple docstring""" _snake_case : Any = ["""auto-generated""", """autogenerated""", """automatically generated"""] _snake_case : Tuple = example["""content"""].splitlines() for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Union[str, Any]=5 , snake_case__ : Any=0.05 ): """simple docstring""" _snake_case : Optional[Any] = ["""unit tests""", """test file""", """configuration file"""] _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : Dict = 0 _snake_case : str = 0 # first test for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _snake_case : Optional[int] = example["""content"""].count("""\n""" ) _snake_case : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[int] = ["""def """, """class """, """for """, """while """] _snake_case : str = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : List[str]=4 ): """simple docstring""" _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : str = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" _snake_case : Optional[Any] = tokenizer(example["""content"""] , truncation=snake_case__ )["""input_ids"""] _snake_case : Optional[Any] = len(example["""content"""] ) / len(snake_case__ ) return {"ratio": ratio} def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = {} results.update(get_hash(snake_case__ ) ) results.update(line_stats(snake_case__ ) ) results.update(alpha_stats(snake_case__ ) ) results.update(char_token_ratio(snake_case__ ) ) results.update(is_autogenerated(snake_case__ ) ) results.update(is_config_or_test(snake_case__ ) ) results.update(has_no_keywords(snake_case__ ) ) results.update(has_few_assignments(snake_case__ ) ) return results def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] ): """simple docstring""" if not check_uniques(snake_case__ , snake_case__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """rb""" ) as f_in: with gzip.open(str(snake_case__ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(snake_case__ , snake_case__ ) os.unlink(snake_case__ ) # Settings A_ = HfArgumentParser(PreprocessingArguments) A_ = parser.parse_args() if args.num_workers is None: A_ = multiprocessing.cpu_count() A_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ = time.time() A_ = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ = time.time() A_ = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ = set(ds.unique('''hash''')) A_ = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ = time.time() A_ = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ = time.time() A_ , A_ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) A_ = output_dir / '''data''' data_dir.mkdir(exist_ok=True) A_ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ = str(data_dir / F'''file-{file_number+1:012}.json''') A_ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int , snake_case__ : float , snake_case__ : float ): """simple docstring""" return round(float(moles / volume ) * nfactor ) def UpperCAmelCase__ (snake_case__ : float , snake_case__ : float , snake_case__ : float ): """simple docstring""" return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def UpperCAmelCase__ (snake_case__ : float , snake_case__ : float , snake_case__ : float ): """simple docstring""" return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def UpperCAmelCase__ (snake_case__ : float , snake_case__ : float , snake_case__ : float ): """simple docstring""" return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
1
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" _snake_case : list[list[int]] = [] _snake_case : list[int] = [] _snake_case : Tuple = 0 _snake_case : Optional[Any] = sum(snake_case__ ) create_state_space_tree(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) return result def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int , snake_case__ : int , snake_case__ : list[int] , snake_case__ : list[list[int]] , snake_case__ : int , ): """simple docstring""" if sum(snake_case__ ) > max_sum or (remaining_nums_sum + sum(snake_case__ )) < max_sum: return if sum(snake_case__ ) == max_sum: result.append(snake_case__ ) return for index in range(snake_case__ , len(snake_case__ ) ): create_state_space_tree( snake_case__ , snake_case__ , index + 1 , [*path, nums[index]] , snake_case__ , remaining_nums_sum - nums[index] , ) A_ = [3, 34, 4, 12, 5, 2] A_ = 9 A_ = generate_sum_of_subsets_soln(nums, max_sum) print(*result)
64
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : str , snake_case__ : str ): """simple docstring""" for attribute in key.split(""".""" ): _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ).shape else: _snake_case : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": _snake_case : int = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : Tuple = value elif weight_type == "bias": _snake_case : List[str] = value else: _snake_case : int = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str] ): """simple docstring""" _snake_case : List[Any] = [] _snake_case : Optional[Any] = fairseq_model.state_dict() _snake_case : str = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _snake_case : Optional[Any] = None for name, value in fairseq_dict.items(): _snake_case : Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) _snake_case : Dict = True elif name.split(""".""" )[0] == "proj": _snake_case : Dict = fairseq_model.proj _snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: _snake_case : Dict = True if "*" in mapped_key: _snake_case : Optional[int] = name.split(snake_case__ )[0].split(""".""" )[-2] _snake_case : Union[str, Any] = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: _snake_case : str = """weight_g""" elif "weight_v" in name: _snake_case : Optional[Any] = """weight_v""" elif "bias" in name: _snake_case : Union[str, Any] = """bias""" elif "weight" in name: _snake_case : int = """weight""" else: _snake_case : Optional[int] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Dict , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] , snake_case__ : int ): """simple docstring""" _snake_case : Any = full_name.split("""conv_layers.""" )[-1] _snake_case : Optional[int] = name.split(""".""" ) _snake_case : List[str] = int(items[0] ) _snake_case : Dict = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _snake_case : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _snake_case : int = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _snake_case : List[str] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" _snake_case , _snake_case : Optional[Any] = emb.weight.shape _snake_case : Optional[int] = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) _snake_case : Union[str, Any] = emb.weight.data return lin_layer def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Any = f.readlines() _snake_case : Optional[Any] = [line.split(""" """ )[0] for line in lines] _snake_case : str = len(snake_case__ ) _snake_case : Tuple = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Union[str, Any] , ): """simple docstring""" _snake_case : Optional[int] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : List[str] = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) _snake_case : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) _snake_case , _snake_case , _snake_case : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) _snake_case : Optional[Any] = model[0].eval() # set weights for wav2vec2 encoder _snake_case : Any = WavaVecaModel(snake_case__ ) _snake_case : Optional[Any] = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) _snake_case : Optional[Any] = SpeechaTextaForCausalLM(snake_case__ ) _snake_case , _snake_case : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) _snake_case : Any = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) _snake_case : Any = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) _snake_case : Any = False # add projection layer _snake_case : int = nn.Parameter(projection_layer.weight ) _snake_case : Any = nn.Parameter(projection_layer.bias ) _snake_case : Any = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) _snake_case : Dict = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) _snake_case : str = hf_wavavec.config.to_dict() _snake_case : List[str] = tokenizer.pad_token_id _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Union[str, Any] = tokenizer.eos_token_id _snake_case : Optional[Any] = """speech_to_text_2""" _snake_case : Optional[int] = """wav2vec2""" _snake_case : Tuple = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_02_24, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') A_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
64
1
"""simple docstring""" A_ = { '''Pillow''': '''Pillow<10.0.0''', '''accelerate''': '''accelerate>=0.20.3''', '''av''': '''av==9.2.0''', '''beautifulsoup4''': '''beautifulsoup4''', '''black''': '''black~=23.1''', '''codecarbon''': '''codecarbon==1.2.0''', '''cookiecutter''': '''cookiecutter==1.7.3''', '''dataclasses''': '''dataclasses''', '''datasets''': '''datasets!=2.5.0''', '''decord''': '''decord==0.6.0''', '''deepspeed''': '''deepspeed>=0.9.3''', '''diffusers''': '''diffusers''', '''dill''': '''dill<0.3.5''', '''evaluate''': '''evaluate>=0.2.0''', '''fairscale''': '''fairscale>0.3''', '''faiss-cpu''': '''faiss-cpu''', '''fastapi''': '''fastapi''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1,<=0.7.0''', '''ftfy''': '''ftfy''', '''fugashi''': '''fugashi>=1.0''', '''GitPython''': '''GitPython<3.1.19''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.14.1,<1.0''', '''importlib_metadata''': '''importlib_metadata''', '''ipadic''': '''ipadic>=1.0.0,<2.0''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2,<=0.4.13''', '''jaxlib''': '''jaxlib>=0.1.65,<=0.4.13''', '''jieba''': '''jieba''', '''kenlm''': '''kenlm''', '''keras-nlp''': '''keras-nlp>=0.3.1''', '''librosa''': '''librosa''', '''nltk''': '''nltk''', '''natten''': '''natten>=0.14.6''', '''numpy''': '''numpy>=1.17''', '''onnxconverter-common''': '''onnxconverter-common''', '''onnxruntime-tools''': '''onnxruntime-tools>=1.4.2''', '''onnxruntime''': '''onnxruntime>=1.4.0''', '''opencv-python''': '''opencv-python''', '''optuna''': '''optuna''', '''optax''': '''optax>=0.0.8,<=0.1.4''', '''packaging''': '''packaging>=20.0''', '''parameterized''': '''parameterized''', '''phonemizer''': '''phonemizer''', '''protobuf''': '''protobuf''', '''psutil''': '''psutil''', '''pyyaml''': '''pyyaml>=5.1''', '''pydantic''': '''pydantic<2''', '''pytest''': '''pytest>=7.2.0''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''python''': '''python>=3.8.0''', '''ray[tune]''': '''ray[tune]''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''rhoknp''': '''rhoknp>=1.1.0,<1.3.1''', '''rjieba''': '''rjieba''', '''rouge-score''': '''rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1''', '''ruff''': '''ruff>=0.0.241,<=0.0.259''', '''sacrebleu''': '''sacrebleu>=1.4.12,<2.0.0''', '''sacremoses''': '''sacremoses''', '''safetensors''': '''safetensors>=0.3.1''', '''sagemaker''': '''sagemaker>=2.31.0''', '''scikit-learn''': '''scikit-learn''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''sigopt''': '''sigopt''', '''starlette''': '''starlette''', '''sudachipy''': '''sudachipy>=0.6.6''', '''sudachidict_core''': '''sudachidict_core>=20220729''', '''tensorflow-cpu''': '''tensorflow-cpu>=2.6,<2.14''', '''tensorflow''': '''tensorflow>=2.6,<2.14''', '''tensorflow-text''': '''tensorflow-text<2.14''', '''tf2onnx''': '''tf2onnx''', '''timeout-decorator''': '''timeout-decorator''', '''timm''': '''timm''', '''tokenizers''': '''tokenizers>=0.11.1,!=0.11.3,<0.14''', '''torch''': '''torch>=1.9,!=1.12.0''', '''torchaudio''': '''torchaudio''', '''torchvision''': '''torchvision''', '''pyctcdecode''': '''pyctcdecode>=0.4.0''', '''tqdm''': '''tqdm>=4.27''', '''unidic''': '''unidic>=1.0.2''', '''unidic_lite''': '''unidic_lite>=1.0.7''', '''urllib3''': '''urllib3<2.0.0''', '''uvicorn''': '''uvicorn''', }
64
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 1_00 ): """simple docstring""" _snake_case : List[Any] = set() _snake_case : Any = 0 _snake_case : List[str] = n + 1 # maximum limit for a in range(2 , snake_case__ ): for b in range(2 , snake_case__ ): _snake_case : int = a**b # calculates the current power collect_powers.add(snake_case__ ) # adds the result to the set return len(snake_case__ ) if __name__ == "__main__": print('''Number of terms ''', solution(int(str(input()).strip())))
64
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A_ = logging.get_logger(__name__) class lowercase( __a ): '''simple docstring''' lowercase__ = ["input_features"] def __init__( self: List[Any], a_: Tuple=80, a_: List[str]=16_000, a_: Tuple=160, a_: Optional[int]=30, a_: List[str]=400, a_: Union[str, Any]=0.0, a_: Union[str, Any]=False, **a_: Dict, ): '''simple docstring''' super().__init__( feature_size=a_, sampling_rate=a_, padding_value=a_, return_attention_mask=a_, **a_, ) _snake_case : Union[str, Any] = n_fft _snake_case : Optional[Any] = hop_length _snake_case : Optional[int] = chunk_length _snake_case : str = chunk_length * sampling_rate _snake_case : Dict = self.n_samples // hop_length _snake_case : List[str] = sampling_rate _snake_case : Optional[int] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2, num_mel_filters=a_, min_frequency=0.0, max_frequency=8_000.0, sampling_rate=a_, norm="""slaney""", mel_scale="""slaney""", ) def UpperCamelCase_ ( self: List[Any], a_: np.array ): '''simple docstring''' _snake_case : str = spectrogram( a_, window_function(self.n_fft, """hann""" ), frame_length=self.n_fft, hop_length=self.hop_length, power=2.0, mel_filters=self.mel_filters, log_mel="""log10""", ) _snake_case : Dict = log_spec[:, :-1] _snake_case : List[Any] = np.maximum(a_, log_spec.max() - 8.0 ) _snake_case : int = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCamelCase_ ( a_: List[np.ndarray], a_: List[np.ndarray], a_: float = 0.0 ): '''simple docstring''' if attention_mask is not None: _snake_case : str = np.array(a_, np.intaa ) _snake_case : str = [] for vector, length in zip(a_, attention_mask.sum(-1 ) ): _snake_case : List[str] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _snake_case : Any = padding_value normed_input_values.append(a_ ) else: _snake_case : Dict = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self: Optional[int], a_: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]], a_: bool = True, a_: Optional[int] = None, a_: Optional[Union[str, TensorType]] = None, a_: Optional[bool] = None, a_: Optional[str] = "max_length", a_: Optional[int] = None, a_: Optional[int] = None, a_: Optional[bool] = None, **a_: Any, ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) _snake_case : str = isinstance(a_, np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) _snake_case : Optional[Any] = is_batched_numpy or ( isinstance(a_, (list, tuple) ) and (isinstance(raw_speech[0], (np.ndarray, tuple, list) )) ) if is_batched: _snake_case : Union[str, Any] = [np.asarray([speech], dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a_, np.ndarray ): _snake_case : Optional[Any] = np.asarray(a_, dtype=np.floataa ) elif isinstance(a_, np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _snake_case : Tuple = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _snake_case : List[str] = [np.asarray([raw_speech] ).T] _snake_case : Optional[int] = BatchFeature({"""input_features""": raw_speech} ) # convert into correct format for padding _snake_case : Optional[Any] = self.pad( a_, padding=a_, max_length=max_length if max_length else self.n_samples, truncation=a_, pad_to_multiple_of=a_, return_attention_mask=return_attention_mask or do_normalize, ) # zero-mean and unit-variance normalization if do_normalize: _snake_case : Dict = self.zero_mean_unit_var_norm( padded_inputs["""input_features"""], attention_mask=padded_inputs["""attention_mask"""], padding_value=self.padding_value, ) _snake_case : Tuple = np.stack(padded_inputs["""input_features"""], axis=0 ) # make sure list is in array format _snake_case : Optional[int] = padded_inputs.get("""input_features""" ).transpose(2, 0, 1 ) _snake_case : Dict = [self._np_extract_fbank_features(a_ ) for waveform in input_features[0]] if isinstance(input_features[0], a_ ): _snake_case : List[str] = [np.asarray(a_, dtype=np.floataa ) for feature in input_features] else: _snake_case : List[Any] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _snake_case : Any = padded_inputs["""attention_mask"""][:, :: self.hop_length] if return_tensors is not None: _snake_case : List[str] = padded_inputs.convert_to_tensors(a_ ) return padded_inputs def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = copy.deepcopy(self.__dict__ ) _snake_case : Union[str, Any] = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
64
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
1
"""simple docstring""" import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging A_ = logging.get_logger(__name__) class lowercase( __a ): '''simple docstring''' lowercase__ = ["input_values", "attention_mask"] def __init__( self: Union[str, Any], a_: int = 1, a_: int = 16_000, a_: float = 0.0, a_: bool = False, a_: int = 80, a_: int = 16, a_: int = 64, a_: str = "hann_window", a_: float = 1.0, a_: float = 80, a_: float = 7_600, a_: float = 1E-10, a_: int = 2, a_: bool = True, **a_: Optional[Any], ): '''simple docstring''' super().__init__(feature_size=a_, sampling_rate=a_, padding_value=a_, **a_ ) _snake_case : Any = do_normalize _snake_case : Optional[int] = return_attention_mask _snake_case : Optional[Any] = num_mel_bins _snake_case : Optional[Any] = hop_length _snake_case : List[Any] = win_length _snake_case : int = win_function _snake_case : Dict = frame_signal_scale _snake_case : Union[str, Any] = fmin _snake_case : Union[str, Any] = fmax _snake_case : Optional[Any] = mel_floor _snake_case : Optional[int] = reduction_factor _snake_case : List[str] = win_length * sampling_rate // 1_000 _snake_case : str = hop_length * sampling_rate // 1_000 _snake_case : Optional[int] = optimal_fft_length(self.sample_size ) _snake_case : List[str] = (self.n_fft // 2) + 1 _snake_case : int = window_function(window_length=self.sample_size, name=self.win_function, periodic=a_ ) _snake_case : Optional[Any] = mel_filter_bank( num_frequency_bins=self.n_freqs, num_mel_filters=self.num_mel_bins, min_frequency=self.fmin, max_frequency=self.fmax, sampling_rate=self.sampling_rate, norm="""slaney""", mel_scale="""slaney""", ) if frame_signal_scale != 1.0: warnings.warn( """The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers""", a_, ) if reduction_factor != 2.0: warnings.warn( """The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers""", a_, ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCamelCase_ ( a_: List[np.ndarray], a_: List[np.ndarray], a_: float = 0.0 ): '''simple docstring''' if attention_mask is not None: _snake_case : Union[str, Any] = np.array(a_, np.intaa ) _snake_case : Optional[int] = [] for vector, length in zip(a_, attention_mask.sum(-1 ) ): _snake_case : str = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _snake_case : Union[str, Any] = padding_value normed_input_values.append(a_ ) else: _snake_case : List[str] = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def UpperCamelCase_ ( self: str, a_: np.ndarray, ): '''simple docstring''' _snake_case : Tuple = spectrogram( a_, window=self.window, frame_length=self.sample_size, hop_length=self.sample_stride, fft_length=self.n_fft, mel_filters=self.mel_filters, mel_floor=self.mel_floor, log_mel="""log10""", ) return log_mel_spec.T def __call__( self: Tuple, a_: Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None, a_: Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None, a_: Union[bool, str, PaddingStrategy] = False, a_: Optional[int] = None, a_: bool = False, a_: Optional[int] = None, a_: Optional[bool] = None, a_: Optional[Union[str, TensorType]] = None, a_: Optional[int] = None, **a_: Any, ): '''simple docstring''' if audio is None and audio_target is None: raise ValueError("""You must provide either `audio` or `audio_target` values.""" ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided audio input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) if audio is not None: _snake_case : Optional[Any] = self._process_audio( a_, a_, a_, a_, a_, a_, a_, a_, **a_, ) else: _snake_case : Optional[int] = None if audio_target is not None: _snake_case : List[Any] = self._process_audio( a_, a_, a_, a_, a_, a_, a_, a_, **a_, ) if inputs is None: return inputs_target else: _snake_case : Tuple = inputs_target["""input_values"""] _snake_case : Any = inputs_target.get("""attention_mask""" ) if decoder_attention_mask is not None: _snake_case : Optional[Any] = decoder_attention_mask return inputs def UpperCamelCase_ ( self: Dict, a_: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]], a_: bool = False, a_: Union[bool, str, PaddingStrategy] = False, a_: Optional[int] = None, a_: bool = False, a_: Optional[int] = None, a_: Optional[bool] = None, a_: Optional[Union[str, TensorType]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = isinstance(a_, np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) _snake_case : Optional[int] = is_batched_numpy or ( isinstance(a_, (list, tuple) ) and (isinstance(speech[0], (np.ndarray, tuple, list) )) ) if is_batched: _snake_case : Tuple = [np.asarray(a_, dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(a_, np.ndarray ): _snake_case : Union[str, Any] = np.asarray(a_, dtype=np.floataa ) elif isinstance(a_, np.ndarray ) and speech.dtype is np.dtype(np.floataa ): _snake_case : Union[str, Any] = speech.astype(np.floataa ) # always return batch if not is_batched: _snake_case : Optional[int] = [speech] # needed to make pad() work on spectrogram inputs _snake_case : Any = self.feature_size # convert into correct format for padding if is_target: _snake_case : str = [self._extract_mel_features(a_ ) for waveform in speech] _snake_case : Dict = BatchFeature({"""input_values""": features} ) _snake_case : Tuple = self.num_mel_bins else: _snake_case : Tuple = BatchFeature({"""input_values""": speech} ) _snake_case : Union[str, Any] = self.pad( a_, padding=a_, max_length=a_, truncation=a_, pad_to_multiple_of=a_, return_attention_mask=a_, **a_, ) _snake_case : int = feature_size_hack # convert input values to correct format _snake_case : List[str] = padded_inputs["""input_values"""] if not isinstance(input_values[0], np.ndarray ): _snake_case : Optional[Any] = [np.asarray(a_, dtype=np.floataa ) for array in input_values] elif ( not isinstance(a_, np.ndarray ) and isinstance(input_values[0], np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): _snake_case : str = [array.astype(np.floataa ) for array in input_values] elif isinstance(a_, np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): _snake_case : Optional[Any] = input_values.astype(np.floataa ) # convert attention_mask to correct format _snake_case : Optional[Any] = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: _snake_case : Dict = [np.asarray(a_, dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: _snake_case : str = ( attention_mask if self._get_padding_strategies(a_, max_length=a_ ) is not PaddingStrategy.DO_NOT_PAD else None ) _snake_case : List[str] = self.zero_mean_unit_var_norm( padded_inputs["""input_values"""], attention_mask=a_, padding_value=self.padding_value ) if return_tensors is not None: _snake_case : Union[str, Any] = padded_inputs.convert_to_tensors(a_ ) return padded_inputs def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : List[Any] = super().to_dict() # Don't serialize these as they are derived from the other properties. _snake_case : List[str] = ["""window""", """mel_filters""", """sample_size""", """sample_stride""", """n_fft""", """n_freqs"""] for name in names: if name in output: del output[name] return output
64
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
1
"""simple docstring""" import cmath import math def UpperCAmelCase__ (snake_case__ : float , snake_case__ : float , snake_case__ : float , snake_case__ : float ): """simple docstring""" _snake_case : Dict = math.radians(snake_case__ ) _snake_case : Dict = math.radians(snake_case__ ) # Convert voltage and current to rectangular form _snake_case : str = cmath.rect(snake_case__ , snake_case__ ) _snake_case : str = cmath.rect(snake_case__ , snake_case__ ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if a < 0: raise ValueError("""Input value must be a positive integer""" ) elif isinstance(snake_case__ , snake_case__ ): raise TypeError("""Input value must be a 'int' type""" ) return bin(snake_case__ ).count("""1""" ) if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = CodeGenTokenizer def __init__( self: Union[str, Any], a_: List[Any]=None, a_: str=None, a_: str=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: str="<|endoftext|>", a_: List[Any]=False, **a_: List[str], ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) if kwargs.pop("""add_bos_token""", a_ ): _snake_case : str = kwargs.pop("""name_or_path""", """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" f"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" f"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) _snake_case : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Dict = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Dict = add_prefix_space _snake_case : str = pre_tok_class(**a_ ) _snake_case : List[Any] = add_prefix_space def UpperCamelCase_ ( self: Any, *a_: Any, **a_: int ): '''simple docstring''' _snake_case : Optional[int] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Any, **a_: List[str] ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: str, a_: Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"], a_: bool = False, a_: bool = None, a_: Optional[List[str]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = super().decode( token_ids=a_, skip_special_tokens=a_, clean_up_tokenization_spaces=a_, **a_, ) if truncate_before_pattern is not None and len(a_ ) > 0: _snake_case : List[str] = self.truncate(a_, a_ ) return decoded_text def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Optional[Any] ): '''simple docstring''' def find_re(a_: Dict, a_: str, a_: Union[str, Any] ): _snake_case : Any = pattern.search(a_, a_ ) return m.start() if m else -1 _snake_case : Tuple = [re.compile(a_, re.MULTILINE ) for pattern in truncate_before_pattern] _snake_case : List[Any] = list(re.finditer("""^print""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : int = completion[: prints[1].start()] _snake_case : List[str] = list(re.finditer("""^def""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : List[Any] = completion[: defs[1].start()] _snake_case : int = 0 _snake_case : List[Any] = [ pos for pos in [find_re(a_, a_, a_ ) for terminal in terminals] if pos != -1 ] if len(a_ ) > 0: return completion[: min(a_ )] else: return completion
64
1
"""simple docstring""" import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class lowercase( __a ): '''simple docstring''' def __init__( self: str, a_: str = "▁", a_: bool = True, a_: Union[str, AddedToken] = "<unk>", a_: Union[str, AddedToken] = "</s>", a_: Union[str, AddedToken] = "<pad>", ): '''simple docstring''' _snake_case : Optional[Any] = { """pad""": {"""id""": 0, """token""": pad_token}, """eos""": {"""id""": 1, """token""": eos_token}, """unk""": {"""id""": 2, """token""": unk_token}, } _snake_case : List[str] = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): _snake_case : Optional[Any] = token_dict["""token"""] _snake_case : Tuple = Tokenizer(Unigram() ) _snake_case : int = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(""" {2,}""" ), """ """ ), normalizers.Lowercase(), ] ) _snake_case : Any = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=a_, add_prefix_space=a_ ), pre_tokenizers.Digits(individual_digits=a_ ), pre_tokenizers.Punctuation(), ] ) _snake_case : Any = decoders.Metaspace(replacement=a_, add_prefix_space=a_ ) _snake_case : List[Any] = TemplateProcessing( single=f"$A {self.special_tokens['eos']['token']}", special_tokens=[(self.special_tokens["""eos"""]["""token"""], self.special_tokens["""eos"""]["""id"""])], ) _snake_case : int = { """model""": """SentencePieceUnigram""", """replacement""": replacement, """add_prefix_space""": add_prefix_space, } super().__init__(a_, a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Union[str, List[str]], a_: int = 8_000, a_: bool = True, ): '''simple docstring''' _snake_case : Dict = trainers.UnigramTrainer( vocab_size=a_, special_tokens=self.special_tokens_list, show_progress=a_, ) if isinstance(a_, a_ ): _snake_case : str = [files] self._tokenizer.train(a_, trainer=a_ ) self.add_unk_id() def UpperCamelCase_ ( self: str, a_: Union[Iterator[str], Iterator[Iterator[str]]], a_: int = 8_000, a_: bool = True, ): '''simple docstring''' _snake_case : Optional[int] = trainers.UnigramTrainer( vocab_size=a_, special_tokens=self.special_tokens_list, show_progress=a_, ) self._tokenizer.train_from_iterator(a_, trainer=a_ ) self.add_unk_id() def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : int = json.loads(self._tokenizer.to_str() ) _snake_case : Tuple = self.special_tokens["""unk"""]["""id"""] _snake_case : Dict = Tokenizer.from_str(json.dumps(a_ ) )
64
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
1
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str]=False ): """simple docstring""" _snake_case : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Dict , snake_case__ : List[str]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : List[Any] = """""" else: _snake_case : List[Any] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _snake_case : Union[str, Any] = in_proj_bias[: config.hidden_size] _snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : List[str] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" _snake_case : List[str] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : int ): """simple docstring""" _snake_case : Optional[Any] = dct.pop(snake_case__ ) _snake_case : Union[str, Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : str = ViTMSNConfig() _snake_case : Any = 10_00 _snake_case : Tuple = """datasets/huggingface/label-files""" _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : int = json.load(open(hf_hub_download(snake_case__ , snake_case__ ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[Any] = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _snake_case : Tuple = 3_84 _snake_case : Dict = 15_36 _snake_case : Tuple = 6 elif "l16" in checkpoint_url: _snake_case : Any = 10_24 _snake_case : int = 40_96 _snake_case : str = 24 _snake_case : Optional[int] = 16 _snake_case : List[Any] = 0.1 elif "b4" in checkpoint_url: _snake_case : Tuple = 4 elif "l7" in checkpoint_url: _snake_case : int = 7 _snake_case : Dict = 10_24 _snake_case : Optional[Any] = 40_96 _snake_case : Any = 24 _snake_case : Union[str, Any] = 16 _snake_case : Optional[int] = 0.1 _snake_case : int = ViTMSNModel(snake_case__ ) _snake_case : Optional[int] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" )["""target_encoder"""] _snake_case : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(snake_case__ ) _snake_case : List[str] = create_rename_keys(snake_case__ , base_model=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) read_in_q_k_v(snake_case__ , snake_case__ , base_model=snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Tuple = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) _snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=snake_case__ , image_std=snake_case__ ) _snake_case : Any = image_processor(images=snake_case__ , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) _snake_case : int = model(**snake_case__ ) _snake_case : List[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _snake_case : Optional[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: _snake_case : str = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: _snake_case : Optional[int] = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: _snake_case : List[Any] = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: _snake_case : Optional[int] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , snake_case__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 10_00 ): """simple docstring""" _snake_case : Tuple = -1 _snake_case : Optional[Any] = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c _snake_case : str = (n * n - 2 * a * n) // (2 * n - 2 * a) _snake_case : Union[str, Any] = n - a - b if c * c == (a * a + b * b): _snake_case : str = a * b * c if candidate >= product: _snake_case : int = candidate return product if __name__ == "__main__": print(F'''{solution() = }''')
64
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''facebook/s2t-wav2vec2-large-en-de''': ( '''https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json''' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowercase( __a ): '''simple docstring''' lowercase__ = "speech_to_text_2" lowercase__ = ["past_key_values"] lowercase__ = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self: List[str], a_: str=10_000, a_: str=6, a_: List[Any]=2_048, a_: Optional[int]=4, a_: List[str]=0.0, a_: Optional[int]=True, a_: List[str]="relu", a_: Optional[Any]=256, a_: Any=0.1, a_: List[str]=0.0, a_: Optional[int]=0.0, a_: Optional[Any]=0.02, a_: Union[str, Any]=2, a_: List[str]=True, a_: Optional[Any]=1, a_: Dict=0, a_: Dict=2, a_: Optional[int]=1_024, **a_: Any, ): '''simple docstring''' _snake_case : Union[str, Any] = vocab_size _snake_case : str = d_model _snake_case : Tuple = decoder_ffn_dim _snake_case : Optional[int] = decoder_layers _snake_case : Optional[Any] = decoder_attention_heads _snake_case : Union[str, Any] = dropout _snake_case : List[str] = attention_dropout _snake_case : str = activation_dropout _snake_case : Union[str, Any] = activation_function _snake_case : List[str] = init_std _snake_case : int = decoder_layerdrop _snake_case : int = use_cache _snake_case : str = decoder_layers _snake_case : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True _snake_case : int = max_target_positions super().__init__( pad_token_id=a_, bos_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, **a_, )
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" stooge(snake_case__ , 0 , len(snake_case__ ) - 1 ) return arr def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any , snake_case__ : int ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) # Recursively sort last 2/3 elements stooge(snake_case__ , i + t , (snake_case__) ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
64
1
"""simple docstring""" import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py A_ = '''src/transformers''' A_ = '''docs/source/en/tasks''' def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : str , snake_case__ : Union[str, Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: _snake_case : Optional[Any] = f.readlines() # Find the start prompt. _snake_case : Union[str, Any] = 0 while not lines[start_index].startswith(snake_case__ ): start_index += 1 start_index += 1 _snake_case : Optional[int] = start_index while not lines[end_index].startswith(snake_case__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. A_ = direct_transformers_import(TRANSFORMERS_PATH) A_ = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). A_ = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = TASK_GUIDE_TO_MODELS[task_guide] _snake_case : Optional[int] = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(snake_case__ , set() ) _snake_case : Dict = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"[{name}](../model_doc/{code})" for code, name in model_names.items()] ) + "\n" def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Optional[int]=False ): """simple docstring""" _snake_case , _snake_case , _snake_case , _snake_case : Optional[Any] = _find_text_in_file( filename=os.path.join(snake_case__ , snake_case__ ) , start_prompt="""<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->""" , end_prompt="""<!--End of the generated tip-->""" , ) _snake_case : Optional[int] = get_model_list_for_task(snake_case__ ) if current_list != new_list: if overwrite: with open(os.path.join(snake_case__ , snake_case__ ) , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`" """ to fix this.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
64
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowercase( metaclass=__a ): '''simple docstring''' lowercase__ = ["note_seq"] def __init__( self: Dict, *a_: Union[str, Any], **a_: List[str] ): '''simple docstring''' requires_backends(self, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Optional[int], *a_: Any, **a_: Optional[Any] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Tuple, *a_: Optional[Any], **a_: List[str] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] )
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 4_00_00_00 ): """simple docstring""" _snake_case : Dict = [0, 1] _snake_case : int = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i += 1 _snake_case : str = 0 for j in range(len(snake_case__ ) - 1 ): if fib[j] % 2 == 0: total += fib[j] return total if __name__ == "__main__": print(F'''{solution() = }''')
64
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: List[str] ): '''simple docstring''' _snake_case : int = data _snake_case : Dict = [0X67452301, 0Xefcdab89, 0X98badcfe, 0X10325476, 0Xc3d2e1f0] @staticmethod def UpperCamelCase_ ( a_: Optional[Any], a_: Dict ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0Xffffffff def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _snake_case : Optional[int] = self.data + padding + struct.pack(""">Q""", 8 * len(self.data ) ) return padded_data def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0, len(self.padded_data ), 64 ) ] def UpperCamelCase_ ( self: Optional[Any], a_: List[Any] ): '''simple docstring''' _snake_case : List[str] = list(struct.unpack(""">16L""", a_ ) ) + [0] * 64 for i in range(16, 80 ): _snake_case : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]), 1 ) return w def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = self.padding() _snake_case : str = self.split_blocks() for block in self.blocks: _snake_case : Any = self.expand_block(a_ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = self.h for i in range(0, 80 ): if 0 <= i < 20: _snake_case : int = (b & c) | ((~b) & d) _snake_case : str = 0X5a827999 elif 20 <= i < 40: _snake_case : Optional[int] = b ^ c ^ d _snake_case : str = 0X6ed9eba1 elif 40 <= i < 60: _snake_case : List[Any] = (b & c) | (b & d) | (c & d) _snake_case : List[Any] = 0X8f1bbcdc elif 60 <= i < 80: _snake_case : List[Any] = b ^ c ^ d _snake_case : int = 0Xca62c1d6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = ( self.rotate(a_, 5 ) + f + e + k + expanded_block[i] & 0Xffffffff, a, self.rotate(a_, 30 ), c, d, ) _snake_case : Union[str, Any] = ( self.h[0] + a & 0Xffffffff, self.h[1] + b & 0Xffffffff, self.h[2] + c & 0Xffffffff, self.h[3] + d & 0Xffffffff, self.h[4] + e & 0Xffffffff, ) return ("{:08x}" * 5).format(*self.h ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = B"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) _snake_case : Union[str, Any] = parser.parse_args() _snake_case : List[Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: _snake_case : str = f.read() else: _snake_case : int = bytes(snake_case__ , """utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
64
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm A_ = logging.get_logger(__name__) @dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = [ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self: Optional[int], **a_: Any ): '''simple docstring''' for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: _snake_case : Dict = deprecated_arg[3:] setattr(self, a_, not kwargs.pop(a_ ) ) logger.warning( f"{deprecated_arg} is depreciated. Please use --no_{positive_arg} or" f" {positive_arg}={kwargs[positive_arg]}" ) _snake_case : List[Any] = kwargs.pop("""torchscript""", self.torchscript ) _snake_case : int = kwargs.pop("""torch_xla_tpu_print_metrics""", self.torch_xla_tpu_print_metrics ) _snake_case : Optional[int] = kwargs.pop("""fp16_opt_level""", self.fpaa_opt_level ) super().__init__(**a_ ) lowercase__ = field(default=__a , metadata={"help": "Trace the models using torchscript"} ) lowercase__ = field(default=__a , metadata={"help": "Print Xla/PyTorch tpu metrics"} ) lowercase__ = field( default="O1" , metadata={ "help": ( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. " "See details at https://nvidia.github.io/apex/amp.html" ) } , ) @cached_property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' requires_backends(self, ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: _snake_case : List[Any] = torch.device("""cpu""" ) _snake_case : int = 0 elif is_torch_tpu_available(): _snake_case : Any = xm.xla_device() _snake_case : int = 0 else: _snake_case : List[str] = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) _snake_case : Dict = torch.cuda.device_count() return device, n_gpu @property def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' return is_torch_tpu_available() and self.tpu @property def UpperCamelCase_ ( self: Dict ): '''simple docstring''' requires_backends(self, ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def UpperCamelCase_ ( self: str ): '''simple docstring''' requires_backends(self, ["""torch"""] ) return self._setup_devices[0] @property def UpperCamelCase_ ( self: Any ): '''simple docstring''' requires_backends(self, ["""torch"""] ) return self._setup_devices[1] @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return self.n_gpu > 0
64
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
1
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str]=False ): """simple docstring""" _snake_case : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Dict , snake_case__ : List[str]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : List[Any] = """""" else: _snake_case : List[Any] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _snake_case : Union[str, Any] = in_proj_bias[: config.hidden_size] _snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : List[str] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" _snake_case : List[str] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : int ): """simple docstring""" _snake_case : Optional[Any] = dct.pop(snake_case__ ) _snake_case : Union[str, Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : str = ViTMSNConfig() _snake_case : Any = 10_00 _snake_case : Tuple = """datasets/huggingface/label-files""" _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : int = json.load(open(hf_hub_download(snake_case__ , snake_case__ ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[Any] = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _snake_case : Tuple = 3_84 _snake_case : Dict = 15_36 _snake_case : Tuple = 6 elif "l16" in checkpoint_url: _snake_case : Any = 10_24 _snake_case : int = 40_96 _snake_case : str = 24 _snake_case : Optional[int] = 16 _snake_case : List[Any] = 0.1 elif "b4" in checkpoint_url: _snake_case : Tuple = 4 elif "l7" in checkpoint_url: _snake_case : int = 7 _snake_case : Dict = 10_24 _snake_case : Optional[Any] = 40_96 _snake_case : Any = 24 _snake_case : Union[str, Any] = 16 _snake_case : Optional[int] = 0.1 _snake_case : int = ViTMSNModel(snake_case__ ) _snake_case : Optional[int] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" )["""target_encoder"""] _snake_case : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(snake_case__ ) _snake_case : List[str] = create_rename_keys(snake_case__ , base_model=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) read_in_q_k_v(snake_case__ , snake_case__ , base_model=snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Tuple = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) _snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=snake_case__ , image_std=snake_case__ ) _snake_case : Any = image_processor(images=snake_case__ , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) _snake_case : int = model(**snake_case__ ) _snake_case : List[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _snake_case : Optional[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: _snake_case : str = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: _snake_case : Optional[int] = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: _snake_case : List[Any] = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: _snake_case : Optional[int] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , snake_case__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
64
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
1
"""simple docstring""" import unittest from dataclasses import dataclass import pytest from accelerate.commands.config.config_args import SageMakerConfig from accelerate.utils import ComputeEnvironment from accelerate.utils.launch import _convert_nargs_to_dict @dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = ComputeEnvironment.AMAZON_SAGEMAKER lowercase__ = True lowercase__ = "ml.p3.2xlarge" lowercase__ = "accelerate_sagemaker_execution_role" lowercase__ = "hf-sm" lowercase__ = "us-east-1" lowercase__ = 1 lowercase__ = "accelerate-sagemaker-1" lowercase__ = "1.6" lowercase__ = "4.4" lowercase__ = "train.py" lowercase__ = [ "--model_name_or_path", "bert", "--do_train", "False", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] lowercase__ = [ "--model_name_or_path", "bert", "--do_train", "--do_test", "False", "--do_predict", "--epochs", "3", "--learning_rate", "5e-5", "--max_steps", "50.5", ] class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[Any] = _convert_nargs_to_dict(MockLaunchConfig.success_training_script_args ) assert isinstance(converted_args["""model_name_or_path"""], a_ ) assert isinstance(converted_args["""do_train"""], a_ ) assert isinstance(converted_args["""epochs"""], a_ ) assert isinstance(converted_args["""learning_rate"""], a_ ) assert isinstance(converted_args["""max_steps"""], a_ ) with pytest.raises(a_ ): _convert_nargs_to_dict(MockLaunchConfig.fail_training_script_args )
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list ): """simple docstring""" if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] _snake_case : List[Any] = [] def generate(snake_case__ : int , snake_case__ : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _snake_case , _snake_case : Optional[Any] = arr[k - 1], arr[i] else: # k is odd _snake_case , _snake_case : List[str] = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
64
1
"""simple docstring""" from collections.abc import Sequence from queue import Queue class lowercase: '''simple docstring''' def __init__( self: Dict, a_: Dict, a_: Any, a_: int, a_: Optional[Any]=None, a_: int=None ): '''simple docstring''' _snake_case : List[str] = start _snake_case : Optional[Any] = end _snake_case : Union[str, Any] = val _snake_case : str = (start + end) // 2 _snake_case : List[str] = left _snake_case : Optional[Any] = right def __repr__( self: int ): '''simple docstring''' return f"SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})" class lowercase: '''simple docstring''' def __init__( self: Union[str, Any], a_: Sequence, a_: Tuple ): '''simple docstring''' _snake_case : List[Any] = collection _snake_case : List[str] = function if self.collection: _snake_case : str = self._build_tree(0, len(a_ ) - 1 ) def UpperCamelCase_ ( self: List[str], a_: Tuple, a_: List[str] ): '''simple docstring''' self._update_tree(self.root, a_, a_ ) def UpperCamelCase_ ( self: Any, a_: List[Any], a_: Optional[Any] ): '''simple docstring''' return self._query_range(self.root, a_, a_ ) def UpperCamelCase_ ( self: Optional[int], a_: List[str], a_: int ): '''simple docstring''' if start == end: return SegmentTreeNode(a_, a_, self.collection[start] ) _snake_case : Optional[Any] = (start + end) // 2 _snake_case : List[str] = self._build_tree(a_, a_ ) _snake_case : List[Any] = self._build_tree(mid + 1, a_ ) return SegmentTreeNode(a_, a_, self.fn(left.val, right.val ), a_, a_ ) def UpperCamelCase_ ( self: Optional[int], a_: List[Any], a_: Tuple, a_: Dict ): '''simple docstring''' if node.start == i and node.end == i: _snake_case : Any = val return if i <= node.mid: self._update_tree(node.left, a_, a_ ) else: self._update_tree(node.right, a_, a_ ) _snake_case : List[Any] = self.fn(node.left.val, node.right.val ) def UpperCamelCase_ ( self: List[str], a_: Any, a_: Any, a_: Dict ): '''simple docstring''' if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left, a_, a_ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left, a_, node.mid ), self._query_range(node.right, node.mid + 1, a_ ), ) else: # range in right child tree return self._query_range(node.right, a_, a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' if self.root is not None: _snake_case : List[str] = Queue() queue.put(self.root ) while not queue.empty(): _snake_case : int = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 50) A_ = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
64
"""simple docstring""" from math import factorial A_ = {str(d): factorial(d) for d in range(10)} def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" return sum(DIGIT_FACTORIAL[d] for d in str(snake_case__ ) ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , snake_case__ ) if sum_of_digit_factorial(snake_case__ ) == i ) if __name__ == "__main__": print(F'''{solution() = }''')
64
1
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" if len(snake_case__ ) < k or k < 0: raise ValueError("""Invalid Input""" ) _snake_case : Optional[int] = sum(array[:k] ) for i in range(len(snake_case__ ) - k ): _snake_case : Optional[Any] = current_sum - array[i] + array[i + k] _snake_case : List[str] = max(snake_case__ , snake_case__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-10_00, 10_00) for i in range(1_00)] A_ = randint(0, 1_10) print(F'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
64
1
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
"""simple docstring""" import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures A_ = logging.get_logger(__name__) @dataclass class lowercase: '''simple docstring''' lowercase__ = field(metadata={"help": "The name of the task to train on: " + ", ".join(glue_processors.keys() )} ) lowercase__ = field( metadata={"help": "The input data dir. Should contain the .tsv files (or other data files) for the task."} ) lowercase__ = field( default=1_28 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowercase__ = field( default=__a , metadata={"help": "Overwrite the cached training and evaluation sets"} ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : int = self.task_name.lower() class lowercase( __a ): '''simple docstring''' lowercase__ = "train" lowercase__ = "dev" lowercase__ = "test" class lowercase( __a ): '''simple docstring''' lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 def __init__( self: str, a_: GlueDataTrainingArguments, a_: PreTrainedTokenizerBase, a_: Optional[int] = None, a_: Union[str, Split] = Split.train, a_: Optional[str] = None, ): '''simple docstring''' warnings.warn( """This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets """ """library. You can have a look at this example script for pointers: """ """https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py""", a_, ) _snake_case : List[str] = args _snake_case : Any = glue_processors[args.task_name]() _snake_case : str = glue_output_modes[args.task_name] if isinstance(a_, a_ ): try: _snake_case : Optional[Any] = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) # Load data features from cache or dataset file _snake_case : Tuple = os.path.join( cache_dir if cache_dir is not None else args.data_dir, f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}", ) _snake_case : Dict = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) _snake_case , _snake_case : int = label_list[2], label_list[1] _snake_case : Optional[int] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _snake_case : Dict = cached_features_file + """.lock""" with FileLock(a_ ): if os.path.exists(a_ ) and not args.overwrite_cache: _snake_case : Optional[Any] = time.time() _snake_case : Tuple = torch.load(a_ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]", time.time() - start ) else: logger.info(f"Creating features from dataset file at {args.data_dir}" ) if mode == Split.dev: _snake_case : str = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: _snake_case : int = self.processor.get_test_examples(args.data_dir ) else: _snake_case : List[Any] = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: _snake_case : Union[str, Any] = examples[:limit_length] _snake_case : Tuple = glue_convert_examples_to_features( a_, a_, max_length=args.max_seq_length, label_list=a_, output_mode=self.output_mode, ) _snake_case : List[str] = time.time() torch.save(self.features, a_ ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self: int ): '''simple docstring''' return len(self.features ) def __getitem__( self: List[str], a_: List[Any] ): '''simple docstring''' return self.features[i] def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return self.label_list
64
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = CodeGenTokenizer lowercase__ = CodeGenTokenizerFast lowercase__ = True lowercase__ = {"add_prefix_space": True} lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : Tuple = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] _snake_case : Tuple = dict(zip(a_, range(len(a_ ) ) ) ) _snake_case : str = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _snake_case : List[Any] = {"""unk_token""": """<unk>"""} _snake_case : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(a_ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(a_ ) ) def UpperCamelCase_ ( self: Any, **a_: int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Any, **a_: str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Dict ): '''simple docstring''' _snake_case : Union[str, Any] = """lower newer""" _snake_case : Tuple = """lower newer""" return input_text, output_text def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = CodeGenTokenizer(self.vocab_file, self.merges_file, **self.special_tokens_map ) _snake_case : Optional[Any] = """lower newer""" _snake_case : Optional[int] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _snake_case : int = tokenizer.tokenize(a_, add_prefix_space=a_ ) self.assertListEqual(a_, a_ ) _snake_case : str = tokens + [tokenizer.unk_token] _snake_case : Optional[int] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : int = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : Dict = """lower newer""" # Testing tokenization _snake_case : Dict = tokenizer.tokenize(a_, add_prefix_space=a_ ) _snake_case : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids without special tokens _snake_case : Optional[Any] = tokenizer.encode(a_, add_special_tokens=a_, add_prefix_space=a_ ) _snake_case : Tuple = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids with special tokens _snake_case : Tuple = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : int = tokenizer.encode(a_, add_prefix_space=a_ ) _snake_case : Optional[Any] = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) # Testing the unknown token _snake_case : Tuple = tokens + [rust_tokenizer.unk_token] _snake_case : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: List[Any]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) # Simple input _snake_case : Any = """This is a simple input""" _snake_case : Optional[int] = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") _snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) # Pair input self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname, pad_token="""<pad>""" ) # Simple input _snake_case : List[Any] = """This is a simple input""" _snake_case : int = ["""This is a simple input looooooooong""", """This is a simple input"""] _snake_case : Any = ("""This is a simple input""", """This is a pair""") _snake_case : str = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] _snake_case : str = tokenizer.pad_token_id _snake_case : Optional[int] = tokenizer(a_, padding="""max_length""", max_length=30, return_tensors="""np""" ) _snake_case : Dict = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) _snake_case : Tuple = tokenizer(*a_, padding="""max_length""", max_length=60, return_tensors="""np""" ) _snake_case : Optional[Any] = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1], 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1], 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1], 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1], 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = """$$$""" _snake_case : List[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname, bos_token=a_, add_bos_token=a_ ) _snake_case : str = """This is a simple input""" _snake_case : int = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Tuple = tokenizer(a_ ) _snake_case : Optional[Any] = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0], a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _snake_case : Optional[int] = tokenizer.decode(out_s.input_ids ) _snake_case : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0], a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) _snake_case : Dict = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" _snake_case : Union[str, Any] = """\nif len_a > len_b: result = a\nelse: result = b""" _snake_case : Optional[Any] = tokenizer.encode(a_ ) _snake_case : Dict = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] _snake_case : Optional[Any] = tokenizer.decode(a_, truncate_before_pattern=a_ ) self.assertEqual(a_, a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass
64
1
"""simple docstring""" import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class lowercase( __a , __a , unittest.TestCase ): '''simple docstring''' lowercase__ = AutoencoderKL lowercase__ = "sample" lowercase__ = 1e-2 @property def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : List[Any] = 4 _snake_case : Tuple = 3 _snake_case : Dict = (32, 32) _snake_case : Optional[int] = floats_tensor((batch_size, num_channels) + sizes ).to(a_ ) return {"sample": image} @property def UpperCamelCase_ ( self: str ): '''simple docstring''' return (3, 32, 32) @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return (3, 32, 32) def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Tuple = { """block_out_channels""": [32, 64], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 4, } _snake_case : Dict = self.dummy_input return init_dict, inputs_dict def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' pass def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skipIf(torch_device == """mps""", """Gradient checkpointing skipped on MPS""" ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case , _snake_case : Optional[Any] = self.prepare_init_args_and_inputs_for_common() _snake_case : str = self.model_class(**a_ ) model.to(a_ ) assert not model.is_gradient_checkpointing and model.training _snake_case : List[str] = model(**a_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() _snake_case : str = torch.randn_like(a_ ) _snake_case : int = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing _snake_case : int = self.model_class(**a_ ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(a_ ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training _snake_case : Tuple = model_a(**a_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() _snake_case : int = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) _snake_case : List[Any] = dict(model.named_parameters() ) _snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data, named_params_a[name].grad.data, atol=5E-5 ) ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case , _snake_case : Optional[int] = AutoencoderKL.from_pretrained("""fusing/autoencoder-kl-dummy""", output_loading_info=a_ ) self.assertIsNotNone(a_ ) self.assertEqual(len(loading_info["""missing_keys"""] ), 0 ) model.to(a_ ) _snake_case : List[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : str = AutoencoderKL.from_pretrained("""fusing/autoencoder-kl-dummy""" ) _snake_case : Tuple = model.to(a_ ) model.eval() if torch_device == "mps": _snake_case : int = torch.manual_seed(0 ) else: _snake_case : Tuple = torch.Generator(device=a_ ).manual_seed(0 ) _snake_case : List[str] = torch.randn( 1, model.config.in_channels, model.config.sample_size, model.config.sample_size, generator=torch.manual_seed(0 ), ) _snake_case : Union[str, Any] = image.to(a_ ) with torch.no_grad(): _snake_case : List[str] = model(a_, sample_posterior=a_, generator=a_ ).sample _snake_case : Tuple = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": _snake_case : Any = torch.tensor( [ -4.00_78E-01, -3.83_23E-04, -1.26_81E-01, -1.14_62E-01, 2.00_95E-01, 1.08_93E-01, -8.82_47E-02, -3.03_61E-01, -9.86_44E-03, ] ) elif torch_device == "cpu": _snake_case : Dict = torch.tensor( [-0.1_352, 0.0_878, 0.0_419, -0.0_818, -0.1_069, 0.0_688, -0.1_458, -0.4_446, -0.0_026] ) else: _snake_case : List[Any] = torch.tensor( [-0.2_421, 0.4_642, 0.2_507, -0.0_438, 0.0_682, 0.3_160, -0.2_018, -0.0_727, 0.2_485] ) self.assertTrue(torch_all_close(a_, a_, rtol=1E-2 ) ) @slow class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: List[str], a_: List[Any], a_: List[Any] ): '''simple docstring''' return f"gaussian_noise_s={seed}_shape={'_'.join([str(a_ ) for s in shape] )}.npy" def UpperCamelCase_ ( self: str ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self: str, a_: List[str]=0, a_: Tuple=(4, 3, 512, 512), a_: Optional[Any]=False ): '''simple docstring''' _snake_case : str = torch.floataa if fpaa else torch.floataa _snake_case : int = torch.from_numpy(load_hf_numpy(self.get_file_format(a_, a_ ) ) ).to(a_ ).to(a_ ) return image def UpperCamelCase_ ( self: Any, a_: Optional[int]="CompVis/stable-diffusion-v1-4", a_: str=False ): '''simple docstring''' _snake_case : str = """fp16""" if fpaa else None _snake_case : Optional[int] = torch.floataa if fpaa else torch.floataa _snake_case : Union[str, Any] = AutoencoderKL.from_pretrained( a_, subfolder="""vae""", torch_dtype=a_, revision=a_, ) model.to(a_ ).eval() return model def UpperCamelCase_ ( self: Union[str, Any], a_: List[str]=0 ): '''simple docstring''' if torch_device == "mps": return torch.manual_seed(a_ ) return torch.Generator(device=a_ ).manual_seed(a_ ) @parameterized.expand( [ # fmt: off [33, [-0.1_603, 0.9_878, -0.0_495, -0.0_790, -0.2_709, 0.8_375, -0.2_060, -0.0_824], [-0.2_395, 0.0_098, 0.0_102, -0.0_709, -0.2_840, -0.0_274, -0.0_718, -0.1_824]], [47, [-0.2_376, 0.1_168, 0.1_332, -0.4_840, -0.2_508, -0.0_791, -0.0_493, -0.4_089], [0.0_350, 0.0_847, 0.0_467, 0.0_344, -0.0_842, -0.0_547, -0.0_633, -0.1_131]], # fmt: on ] ) def UpperCamelCase_ ( self: Dict, a_: Any, a_: Any, a_: int ): '''simple docstring''' _snake_case : str = self.get_sd_vae_model() _snake_case : str = self.get_sd_image(a_ ) _snake_case : Dict = self.get_generator(a_ ) with torch.no_grad(): _snake_case : str = model(a_, generator=a_, sample_posterior=a_ ).sample assert sample.shape == image.shape _snake_case : Optional[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() _snake_case : Optional[int] = torch.tensor(expected_slice_mps if torch_device == """mps""" else expected_slice ) assert torch_all_close(a_, a_, atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.0_513, 0.0_289, 1.3_799, 0.2_166, -0.2_573, -0.0_871, 0.5_103, -0.0_999]], [47, [-0.4_128, -0.1_320, -0.3_704, 0.1_965, -0.4_116, -0.2_332, -0.3_340, 0.2_247]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase_ ( self: Tuple, a_: int, a_: Dict ): '''simple docstring''' _snake_case : Tuple = self.get_sd_vae_model(fpaa=a_ ) _snake_case : Tuple = self.get_sd_image(a_, fpaa=a_ ) _snake_case : Tuple = self.get_generator(a_ ) with torch.no_grad(): _snake_case : Optional[int] = model(a_, generator=a_, sample_posterior=a_ ).sample assert sample.shape == image.shape _snake_case : Dict = sample[-1, -2:, :2, -2:].flatten().float().cpu() _snake_case : Tuple = torch.tensor(a_ ) assert torch_all_close(a_, a_, atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.1_609, 0.9_866, -0.0_487, -0.0_777, -0.2_716, 0.8_368, -0.2_055, -0.0_814], [-0.2_395, 0.0_098, 0.0_102, -0.0_709, -0.2_840, -0.0_274, -0.0_718, -0.1_824]], [47, [-0.2_377, 0.1_147, 0.1_333, -0.4_841, -0.2_506, -0.0_805, -0.0_491, -0.4_085], [0.0_350, 0.0_847, 0.0_467, 0.0_344, -0.0_842, -0.0_547, -0.0_633, -0.1_131]], # fmt: on ] ) def UpperCamelCase_ ( self: List[str], a_: Tuple, a_: List[Any], a_: int ): '''simple docstring''' _snake_case : List[Any] = self.get_sd_vae_model() _snake_case : Union[str, Any] = self.get_sd_image(a_ ) with torch.no_grad(): _snake_case : Optional[int] = model(a_ ).sample assert sample.shape == image.shape _snake_case : int = sample[-1, -2:, -2:, :2].flatten().float().cpu() _snake_case : Optional[Any] = torch.tensor(expected_slice_mps if torch_device == """mps""" else expected_slice ) assert torch_all_close(a_, a_, atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.2_051, -0.1_803, -0.2_311, -0.2_114, -0.3_292, -0.3_574, -0.2_953, -0.3_323]], [37, [-0.2_632, -0.2_625, -0.2_199, -0.2_741, -0.4_539, -0.4_990, -0.3_720, -0.4_925]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase_ ( self: Union[str, Any], a_: Optional[int], a_: Any ): '''simple docstring''' _snake_case : Union[str, Any] = self.get_sd_vae_model() _snake_case : List[str] = self.get_sd_image(a_, shape=(3, 4, 64, 64) ) with torch.no_grad(): _snake_case : Union[str, Any] = model.decode(a_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] _snake_case : Union[str, Any] = sample[-1, -2:, :2, -2:].flatten().cpu() _snake_case : List[Any] = torch.tensor(a_ ) assert torch_all_close(a_, a_, atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.0_369, 0.0_207, -0.0_776, -0.0_682, -0.1_747, -0.1_930, -0.1_465, -0.2_039]], [16, [-0.1_628, -0.2_134, -0.2_747, -0.2_642, -0.3_774, -0.4_404, -0.3_687, -0.4_277]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase_ ( self: Union[str, Any], a_: Optional[Any], a_: int ): '''simple docstring''' _snake_case : Tuple = self.get_sd_vae_model(fpaa=a_ ) _snake_case : Union[str, Any] = self.get_sd_image(a_, shape=(3, 4, 64, 64), fpaa=a_ ) with torch.no_grad(): _snake_case : Any = model.decode(a_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] _snake_case : int = sample[-1, -2:, :2, -2:].flatten().float().cpu() _snake_case : Tuple = torch.tensor(a_ ) assert torch_all_close(a_, a_, atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available(), reason="""xformers is not required when using PyTorch 2.0.""" ) def UpperCamelCase_ ( self: Any, a_: Optional[Any] ): '''simple docstring''' _snake_case : Dict = self.get_sd_vae_model(fpaa=a_ ) _snake_case : List[Any] = self.get_sd_image(a_, shape=(3, 4, 64, 64), fpaa=a_ ) with torch.no_grad(): _snake_case : Optional[int] = model.decode(a_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): _snake_case : List[str] = model.decode(a_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(a_, a_, atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available(), reason="""xformers is not required when using PyTorch 2.0.""" ) def UpperCamelCase_ ( self: str, a_: Union[str, Any] ): '''simple docstring''' _snake_case : List[Any] = self.get_sd_vae_model() _snake_case : Any = self.get_sd_image(a_, shape=(3, 4, 64, 64) ) with torch.no_grad(): _snake_case : int = model.decode(a_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): _snake_case : Union[str, Any] = model.decode(a_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(a_, a_, atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.3_001, 0.0_918, -2.6_984, -3.9_720, -3.2_099, -5.0_353, 1.7_338, -0.2_065, 3.4_267]], [47, [-1.5_030, -4.3_871, -6.0_355, -9.1_157, -1.6_661, -2.7_853, 2.1_607, -5.0_823, 2.5_633]], # fmt: on ] ) def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Tuple ): '''simple docstring''' _snake_case : str = self.get_sd_vae_model() _snake_case : int = self.get_sd_image(a_ ) _snake_case : Dict = self.get_generator(a_ ) with torch.no_grad(): _snake_case : Dict = model.encode(a_ ).latent_dist _snake_case : Dict = dist.sample(generator=a_ ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] _snake_case : Optional[int] = sample[0, -1, -3:, -3:].flatten().cpu() _snake_case : Tuple = torch.tensor(a_ ) _snake_case : List[Any] = 3E-3 if torch_device != """mps""" else 1E-2 assert torch_all_close(a_, a_, atol=a_ )
64
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ = re.compile(r'''\s+''') def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" return {"hash": hashlib.mda(re.sub(snake_case__ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : Any = [len(snake_case__ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(snake_case__ ), "line_max": max(snake_case__ )} def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : List[Any] ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : List[str]=5 ): """simple docstring""" _snake_case : Any = ["""auto-generated""", """autogenerated""", """automatically generated"""] _snake_case : Tuple = example["""content"""].splitlines() for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Union[str, Any]=5 , snake_case__ : Any=0.05 ): """simple docstring""" _snake_case : Optional[Any] = ["""unit tests""", """test file""", """configuration file"""] _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : Dict = 0 _snake_case : str = 0 # first test for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _snake_case : Optional[int] = example["""content"""].count("""\n""" ) _snake_case : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[int] = ["""def """, """class """, """for """, """while """] _snake_case : str = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : List[str]=4 ): """simple docstring""" _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : str = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" _snake_case : Optional[Any] = tokenizer(example["""content"""] , truncation=snake_case__ )["""input_ids"""] _snake_case : Optional[Any] = len(example["""content"""] ) / len(snake_case__ ) return {"ratio": ratio} def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = {} results.update(get_hash(snake_case__ ) ) results.update(line_stats(snake_case__ ) ) results.update(alpha_stats(snake_case__ ) ) results.update(char_token_ratio(snake_case__ ) ) results.update(is_autogenerated(snake_case__ ) ) results.update(is_config_or_test(snake_case__ ) ) results.update(has_no_keywords(snake_case__ ) ) results.update(has_few_assignments(snake_case__ ) ) return results def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] ): """simple docstring""" if not check_uniques(snake_case__ , snake_case__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """rb""" ) as f_in: with gzip.open(str(snake_case__ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(snake_case__ , snake_case__ ) os.unlink(snake_case__ ) # Settings A_ = HfArgumentParser(PreprocessingArguments) A_ = parser.parse_args() if args.num_workers is None: A_ = multiprocessing.cpu_count() A_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ = time.time() A_ = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ = time.time() A_ = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ = set(ds.unique('''hash''')) A_ = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ = time.time() A_ = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ = time.time() A_ , A_ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) A_ = output_dir / '''data''' data_dir.mkdir(exist_ok=True) A_ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ = str(data_dir / F'''file-{file_number+1:012}.json''') A_ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
64
1
"""simple docstring""" import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features A_ = logging.get_logger(__name__) A_ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) A_ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowercase: '''simple docstring''' lowercase__ = field( default=__a , metadata={"help": "Model type selected in the list: " + ", ".join(__a )} ) lowercase__ = field( default=__a , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) lowercase__ = field( default=1_28 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowercase__ = field( default=1_28 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) lowercase__ = field( default=64 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) lowercase__ = field( default=30 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) lowercase__ = field( default=__a , metadata={"help": "Overwrite the cached training and evaluation sets"} ) lowercase__ = field( default=__a , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) lowercase__ = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) lowercase__ = field( default=20 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) lowercase__ = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) lowercase__ = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class lowercase( __a ): '''simple docstring''' lowercase__ = "train" lowercase__ = "dev" class lowercase( __a ): '''simple docstring''' lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 def __init__( self: Any, a_: SquadDataTrainingArguments, a_: PreTrainedTokenizer, a_: Optional[int] = None, a_: Union[str, Split] = Split.train, a_: Optional[bool] = False, a_: Optional[str] = None, a_: Optional[str] = "pt", ): '''simple docstring''' _snake_case : int = args _snake_case : str = is_language_sensitive _snake_case : str = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(a_, a_ ): try: _snake_case : Any = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) _snake_case : int = mode # Load data features from cache or dataset file _snake_case : str = """v2""" if args.version_2_with_negative else """v1""" _snake_case : Dict = os.path.join( cache_dir if cache_dir is not None else args.data_dir, f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}", ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _snake_case : Any = cached_features_file + """.lock""" with FileLock(a_ ): if os.path.exists(a_ ) and not args.overwrite_cache: _snake_case : Optional[Any] = time.time() _snake_case : Optional[int] = torch.load(a_ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. _snake_case : str = self.old_features["""features"""] _snake_case : Dict = self.old_features.get("""dataset""", a_ ) _snake_case : Union[str, Any] = self.old_features.get("""examples""", a_ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]", time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" """ future run""" ) else: if mode == Split.dev: _snake_case : List[str] = self.processor.get_dev_examples(args.data_dir ) else: _snake_case : Tuple = self.processor.get_train_examples(args.data_dir ) _snake_case , _snake_case : Any = squad_convert_examples_to_features( examples=self.examples, tokenizer=a_, max_seq_length=args.max_seq_length, doc_stride=args.doc_stride, max_query_length=args.max_query_length, is_training=mode == Split.train, threads=args.threads, return_dataset=a_, ) _snake_case : str = time.time() torch.save( {"""features""": self.features, """dataset""": self.dataset, """examples""": self.examples}, a_, ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self: List[Any] ): '''simple docstring''' return len(self.features ) def __getitem__( self: Tuple, a_: str ): '''simple docstring''' _snake_case : Optional[int] = self.features[i] _snake_case : Any = torch.tensor(feature.input_ids, dtype=torch.long ) _snake_case : List[Any] = torch.tensor(feature.attention_mask, dtype=torch.long ) _snake_case : Any = torch.tensor(feature.token_type_ids, dtype=torch.long ) _snake_case : Optional[int] = torch.tensor(feature.cls_index, dtype=torch.long ) _snake_case : Any = torch.tensor(feature.p_mask, dtype=torch.float ) _snake_case : int = torch.tensor(feature.is_impossible, dtype=torch.float ) _snake_case : int = { """input_ids""": input_ids, """attention_mask""": attention_mask, """token_type_ids""": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"""cls_index""": cls_index, """p_mask""": p_mask} ) if self.args.version_2_with_negative: inputs.update({"""is_impossible""": is_impossible} ) if self.is_language_sensitive: inputs.update({"""langs""": (torch.ones(input_ids.shape, dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: _snake_case : List[Any] = torch.tensor(feature.start_position, dtype=torch.long ) _snake_case : List[Any] = torch.tensor(feature.end_position, dtype=torch.long ) inputs.update({"""start_positions""": start_positions, """end_positions""": end_positions} ) return inputs
64
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
64
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : str , snake_case__ : str ): """simple docstring""" for attribute in key.split(""".""" ): _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ).shape else: _snake_case : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": _snake_case : int = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : Tuple = value elif weight_type == "bias": _snake_case : List[str] = value else: _snake_case : int = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str] ): """simple docstring""" _snake_case : List[Any] = [] _snake_case : Optional[Any] = fairseq_model.state_dict() _snake_case : str = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _snake_case : Optional[Any] = None for name, value in fairseq_dict.items(): _snake_case : Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) _snake_case : Dict = True elif name.split(""".""" )[0] == "proj": _snake_case : Dict = fairseq_model.proj _snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: _snake_case : Dict = True if "*" in mapped_key: _snake_case : Optional[int] = name.split(snake_case__ )[0].split(""".""" )[-2] _snake_case : Union[str, Any] = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: _snake_case : str = """weight_g""" elif "weight_v" in name: _snake_case : Optional[Any] = """weight_v""" elif "bias" in name: _snake_case : Union[str, Any] = """bias""" elif "weight" in name: _snake_case : int = """weight""" else: _snake_case : Optional[int] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Dict , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] , snake_case__ : int ): """simple docstring""" _snake_case : Any = full_name.split("""conv_layers.""" )[-1] _snake_case : Optional[int] = name.split(""".""" ) _snake_case : List[str] = int(items[0] ) _snake_case : Dict = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _snake_case : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _snake_case : int = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _snake_case : List[str] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" _snake_case , _snake_case : Optional[Any] = emb.weight.shape _snake_case : Optional[int] = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) _snake_case : Union[str, Any] = emb.weight.data return lin_layer def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Any = f.readlines() _snake_case : Optional[Any] = [line.split(""" """ )[0] for line in lines] _snake_case : str = len(snake_case__ ) _snake_case : Tuple = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Union[str, Any] , ): """simple docstring""" _snake_case : Optional[int] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : List[str] = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) _snake_case : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) _snake_case , _snake_case , _snake_case : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) _snake_case : Optional[Any] = model[0].eval() # set weights for wav2vec2 encoder _snake_case : Any = WavaVecaModel(snake_case__ ) _snake_case : Optional[Any] = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) _snake_case : Optional[Any] = SpeechaTextaForCausalLM(snake_case__ ) _snake_case , _snake_case : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) _snake_case : Any = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) _snake_case : Any = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) _snake_case : Any = False # add projection layer _snake_case : int = nn.Parameter(projection_layer.weight ) _snake_case : Any = nn.Parameter(projection_layer.bias ) _snake_case : Any = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) _snake_case : Dict = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) _snake_case : str = hf_wavavec.config.to_dict() _snake_case : List[str] = tokenizer.pad_token_id _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Union[str, Any] = tokenizer.eos_token_id _snake_case : Optional[Any] = """speech_to_text_2""" _snake_case : Optional[int] = """wav2vec2""" _snake_case : Tuple = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_02_24, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') A_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 10_00 ): """simple docstring""" return sum(e for e in range(3 , snake_case__ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F'''{solution() = }''')
64
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A_ = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
64
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
1
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def UpperCAmelCase__ (): """simple docstring""" _snake_case : Union[str, Any] = [randint(-10_00 , 10_00 ) for i in range(10 )] _snake_case : str = randint(-50_00 , 50_00 ) return (arr, r) A_ = make_dataset() def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" for triplet in permutations(snake_case__ , 3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" arr.sort() _snake_case : Any = len(snake_case__ ) for i in range(n - 1 ): _snake_case , _snake_case : Optional[int] = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Tuple = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _snake_case : Optional[Any] = """ triplet_sum1(*dataset) """ _snake_case : int = """ triplet_sum2(*dataset) """ _snake_case : List[str] = repeat(setup=snake_case__ , stmt=snake_case__ , repeat=5 , number=1_00_00 ) _snake_case : Optional[Any] = repeat(setup=snake_case__ , stmt=snake_case__ , repeat=5 , number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() A_ = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
64
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
1
"""simple docstring""" from __future__ import annotations import numpy as np def UpperCAmelCase__ (snake_case__ : np.ndarray ): """simple docstring""" _snake_case , _snake_case : str = np.shape(snake_case__ ) if rows != columns: _snake_case : Any = ( """'table' has to be of square shaped array but got a """ F"{rows}x{columns} array:\n{table}" ) raise ValueError(snake_case__ ) _snake_case : List[Any] = np.zeros((rows, columns) ) _snake_case : List[Any] = np.zeros((rows, columns) ) for i in range(snake_case__ ): for j in range(snake_case__ ): _snake_case : Optional[Any] = sum(lower[i][k] * upper[k][j] for k in range(snake_case__ ) ) if upper[j][j] == 0: raise ArithmeticError("""No LU decomposition exists""" ) _snake_case : Any = (table[i][j] - total) / upper[j][j] _snake_case : Union[str, Any] = 1 for j in range(snake_case__ , snake_case__ ): _snake_case : Optional[Any] = sum(lower[i][k] * upper[k][j] for k in range(snake_case__ ) ) _snake_case : Tuple = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
1
"""simple docstring""" import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors A_ = logging.getLogger(__name__) class lowercase( __a ): '''simple docstring''' lowercase__ = "sequence-classification" def __init__( self: Tuple, a_: List[Any] ): '''simple docstring''' if type(a_ ) == dict: _snake_case : List[Any] = Namespace(**a_ ) _snake_case : str = glue_output_modes[hparams.task] _snake_case : int = glue_tasks_num_labels[hparams.task] super().__init__(a_, a_, self.mode ) def UpperCamelCase_ ( self: Tuple, **a_: Optional[Any] ): '''simple docstring''' return self.model(**a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Dict ): '''simple docstring''' _snake_case : Optional[int] = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: _snake_case : int = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None _snake_case : Any = self(**a_ ) _snake_case : Any = outputs[0] _snake_case : str = self.trainer.lr_schedulers[0]["""scheduler"""] _snake_case : int = {"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : str = self.hparams _snake_case : int = processors[args.task]() _snake_case : Tuple = processor.get_labels() for mode in ["train", "dev"]: _snake_case : Union[str, Any] = self._feature_file(a_ ) if os.path.exists(a_ ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""", a_ ) else: logger.info("""Creating features from dataset file at %s""", args.data_dir ) _snake_case : Dict = ( processor.get_dev_examples(args.data_dir ) if mode == """dev""" else processor.get_train_examples(args.data_dir ) ) _snake_case : Any = convert_examples_to_features( a_, self.tokenizer, max_length=args.max_seq_length, label_list=self.labels, output_mode=args.glue_output_mode, ) logger.info("""Saving features into cached file %s""", a_ ) torch.save(a_, a_ ) def UpperCamelCase_ ( self: Any, a_: str, a_: int, a_: bool = False ): '''simple docstring''' _snake_case : Dict = """dev""" if mode == """test""" else mode _snake_case : Optional[int] = self._feature_file(a_ ) logger.info("""Loading features from cached file %s""", a_ ) _snake_case : List[Any] = torch.load(a_ ) _snake_case : Optional[int] = torch.tensor([f.input_ids for f in features], dtype=torch.long ) _snake_case : Union[str, Any] = torch.tensor([f.attention_mask for f in features], dtype=torch.long ) _snake_case : Optional[Any] = torch.tensor([f.token_type_ids for f in features], dtype=torch.long ) if self.hparams.glue_output_mode == "classification": _snake_case : Tuple = torch.tensor([f.label for f in features], dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": _snake_case : List[str] = torch.tensor([f.label for f in features], dtype=torch.float ) return DataLoader( TensorDataset(a_, a_, a_, a_ ), batch_size=a_, shuffle=a_, ) def UpperCamelCase_ ( self: Union[str, Any], a_: Tuple, a_: int ): '''simple docstring''' _snake_case : Union[str, Any] = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: _snake_case : str = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None _snake_case : int = self(**a_ ) _snake_case , _snake_case : str = outputs[:2] _snake_case : Tuple = logits.detach().cpu().numpy() _snake_case : Tuple = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase_ ( self: Optional[int], a_: Tuple ): '''simple docstring''' _snake_case : Any = torch.stack([x["""val_loss"""] for x in outputs] ).mean().detach().cpu().item() _snake_case : Optional[int] = np.concatenate([x["""pred"""] for x in outputs], axis=0 ) if self.hparams.glue_output_mode == "classification": _snake_case : Tuple = np.argmax(a_, axis=1 ) elif self.hparams.glue_output_mode == "regression": _snake_case : Optional[Any] = np.squeeze(a_ ) _snake_case : List[Any] = np.concatenate([x["""target"""] for x in outputs], axis=0 ) _snake_case : Optional[Any] = [[] for _ in range(out_label_ids.shape[0] )] _snake_case : Tuple = [[] for _ in range(out_label_ids.shape[0] )] _snake_case : str = {**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task, a_, a_ )} _snake_case : str = dict(results.items() ) _snake_case : Optional[Any] = results return ret, preds_list, out_label_list def UpperCamelCase_ ( self: str, a_: list ): '''simple docstring''' _snake_case , _snake_case , _snake_case : List[Any] = self._eval_end(a_ ) _snake_case : Any = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase_ ( self: Optional[int], a_: str ): '''simple docstring''' _snake_case , _snake_case , _snake_case : Optional[int] = self._eval_end(a_ ) _snake_case : str = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase_ ( a_: Optional[int], a_: List[Any] ): '''simple docstring''' BaseTransformer.add_model_specific_args(a_, a_ ) parser.add_argument( """--max_seq_length""", default=128, type=a_, help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ), ) parser.add_argument( """--task""", default="""""", type=a_, required=a_, help="""The GLUE task to run""", ) parser.add_argument( """--gpus""", default=0, type=a_, help="""The number of GPUs allocated for this, it is by default 0 meaning none""", ) parser.add_argument( """--overwrite_cache""", action="""store_true""", help="""Overwrite the cached training and evaluation sets""" ) return parser def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = argparse.ArgumentParser() add_generic_args(snake_case__ , os.getcwd() ) _snake_case : Any = GLUETransformer.add_model_specific_args(snake_case__ , os.getcwd() ) _snake_case : Optional[Any] = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: _snake_case : Optional[int] = os.path.join( """./results""" , F"{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}" , ) os.makedirs(args.output_dir ) _snake_case : Union[str, Any] = GLUETransformer(snake_case__ ) _snake_case : Dict = generic_train(snake_case__ , snake_case__ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: _snake_case : Optional[int] = sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=snake_case__ ) ) _snake_case : Union[str, Any] = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(snake_case__ ) if __name__ == "__main__": main()
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
1
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = CodeGenTokenizer def __init__( self: Union[str, Any], a_: List[Any]=None, a_: str=None, a_: str=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: str="<|endoftext|>", a_: List[Any]=False, **a_: List[str], ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) if kwargs.pop("""add_bos_token""", a_ ): _snake_case : str = kwargs.pop("""name_or_path""", """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" f"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" f"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) _snake_case : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Dict = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Dict = add_prefix_space _snake_case : str = pre_tok_class(**a_ ) _snake_case : List[Any] = add_prefix_space def UpperCamelCase_ ( self: Any, *a_: Any, **a_: int ): '''simple docstring''' _snake_case : Optional[int] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Any, **a_: List[str] ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: str, a_: Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"], a_: bool = False, a_: bool = None, a_: Optional[List[str]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = super().decode( token_ids=a_, skip_special_tokens=a_, clean_up_tokenization_spaces=a_, **a_, ) if truncate_before_pattern is not None and len(a_ ) > 0: _snake_case : List[str] = self.truncate(a_, a_ ) return decoded_text def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Optional[Any] ): '''simple docstring''' def find_re(a_: Dict, a_: str, a_: Union[str, Any] ): _snake_case : Any = pattern.search(a_, a_ ) return m.start() if m else -1 _snake_case : Tuple = [re.compile(a_, re.MULTILINE ) for pattern in truncate_before_pattern] _snake_case : List[Any] = list(re.finditer("""^print""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : int = completion[: prints[1].start()] _snake_case : List[str] = list(re.finditer("""^def""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : List[Any] = completion[: defs[1].start()] _snake_case : int = 0 _snake_case : List[Any] = [ pos for pos in [find_re(a_, a_, a_ ) for terminal in terminals] if pos != -1 ] if len(a_ ) > 0: return completion[: min(a_ )] else: return completion
64
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''MIT/ast-finetuned-audioset-10-10-0.4593''': ( '''https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json''' ), } class lowercase( __a ): '''simple docstring''' lowercase__ = "audio-spectrogram-transformer" def __init__( self: str, a_: Optional[Any]=768, a_: str=12, a_: Dict=12, a_: List[str]=3_072, a_: str="gelu", a_: Dict=0.0, a_: str=0.0, a_: Any=0.02, a_: Tuple=1E-12, a_: Union[str, Any]=16, a_: Dict=True, a_: Any=10, a_: Optional[Any]=10, a_: Tuple=1_024, a_: Optional[int]=128, **a_: int, ): '''simple docstring''' super().__init__(**a_ ) _snake_case : List[str] = hidden_size _snake_case : Dict = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : str = intermediate_size _snake_case : int = hidden_act _snake_case : str = hidden_dropout_prob _snake_case : str = attention_probs_dropout_prob _snake_case : Optional[Any] = initializer_range _snake_case : List[str] = layer_norm_eps _snake_case : int = patch_size _snake_case : Any = qkv_bias _snake_case : Tuple = frequency_stride _snake_case : List[str] = time_stride _snake_case : Any = max_length _snake_case : int = num_mel_bins
64
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
1
"""simple docstring""" import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html A_ = '''platform''' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : str=None , snake_case__ : Optional[int]=None , snake_case__ : List[Any]=None , snake_case__ : int=None , snake_case__ : Union[str, Any]=None , snake_case__ : Tuple=None , ): """simple docstring""" if attention_mask is None: _snake_case : Dict = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case : Tuple = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case : Tuple = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Dict = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case : Any = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class lowercase: '''simple docstring''' def __init__( self: Dict, a_: Any, a_: str=13, a_: int=7, a_: Any=True, a_: Union[str, Any]=False, a_: str=99, a_: List[Any]=16, a_: Optional[Any]=2, a_: int=4, a_: int=4, a_: Dict="gelu", a_: int=0.1, a_: Dict=0.1, a_: Tuple=32, a_: Optional[int]=2, a_: Any=1, a_: List[str]=0, a_: Union[str, Any]=0.02, ): '''simple docstring''' _snake_case : Optional[int] = parent _snake_case : Tuple = batch_size _snake_case : int = seq_length _snake_case : Optional[int] = is_training _snake_case : Tuple = use_labels _snake_case : List[str] = vocab_size _snake_case : Dict = hidden_size _snake_case : int = num_hidden_layers _snake_case : Tuple = num_attention_heads _snake_case : Optional[int] = intermediate_size _snake_case : Any = hidden_act _snake_case : Optional[int] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : Tuple = max_position_embeddings _snake_case : Any = eos_token_id _snake_case : List[str] = pad_token_id _snake_case : Dict = bos_token_id _snake_case : Tuple = initializer_range def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Any = np.clip(ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ), 3, self.vocab_size ) _snake_case : Tuple = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1), dtype=np.intaa )), -1 ) _snake_case : List[str] = shift_tokens_right(a_, 1, 2 ) _snake_case : List[str] = BlenderbotConfig( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_id=self.eos_token_id, bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, initializer_range=self.initializer_range, use_cache=a_, ) _snake_case : Tuple = prepare_blenderbot_inputs_dict(a_, a_, a_ ) return config, inputs_dict def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case , _snake_case : Tuple = self.prepare_config_and_inputs() return config, inputs_dict def UpperCamelCase_ ( self: Union[str, Any], a_: Optional[Any], a_: Any, a_: Union[str, Any] ): '''simple docstring''' _snake_case : Optional[Any] = 20 _snake_case : List[str] = model_class_name(a_ ) _snake_case : Optional[Any] = model.encode(inputs_dict["""input_ids"""] ) _snake_case , _snake_case : Optional[Any] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case : Tuple = model.init_cache(decoder_input_ids.shape[0], a_, a_ ) _snake_case : Optional[Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length), dtype="""i4""" ) _snake_case : Any = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :], (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1), ) _snake_case : List[str] = model.decode( decoder_input_ids[:, :-1], a_, decoder_attention_mask=a_, past_key_values=a_, decoder_position_ids=a_, ) _snake_case : Union[str, Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]], dtype="""i4""" ) _snake_case : Tuple = model.decode( decoder_input_ids[:, -1:], a_, decoder_attention_mask=a_, past_key_values=outputs_cache.past_key_values, decoder_position_ids=a_, ) _snake_case : List[str] = model.decode(a_, a_ ) _snake_case : Tuple = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3, msg=f"Max diff is {diff}" ) def UpperCamelCase_ ( self: Union[str, Any], a_: Optional[Any], a_: List[Any], a_: Any ): '''simple docstring''' _snake_case : Dict = 20 _snake_case : Optional[int] = model_class_name(a_ ) _snake_case : Optional[Any] = model.encode(inputs_dict["""input_ids"""] ) _snake_case , _snake_case : Dict = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case : Any = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ], axis=-1, ) _snake_case : List[Any] = model.init_cache(decoder_input_ids.shape[0], a_, a_ ) _snake_case : Union[str, Any] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :], (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1), ) _snake_case : str = model.decode( decoder_input_ids[:, :-1], a_, decoder_attention_mask=a_, past_key_values=a_, decoder_position_ids=a_, ) _snake_case : str = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]], dtype="""i4""" ) _snake_case : Optional[int] = model.decode( decoder_input_ids[:, -1:], a_, past_key_values=outputs_cache.past_key_values, decoder_attention_mask=a_, decoder_position_ids=a_, ) _snake_case : Optional[int] = model.decode(a_, a_, decoder_attention_mask=a_ ) _snake_case : Optional[Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3, msg=f"Max diff is {diff}" ) @require_flax class lowercase( unittest.TestCase ): '''simple docstring''' lowercase__ = 99 def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : List[Any] = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ], dtype=np.intaa, ) _snake_case : Any = input_ids.shape[0] _snake_case : int = BlenderbotConfig( vocab_size=self.vocab_size, d_model=24, encoder_layers=2, decoder_layers=2, encoder_attention_heads=2, decoder_attention_heads=2, encoder_ffn_dim=32, decoder_ffn_dim=32, max_position_embeddings=48, eos_token_id=2, pad_token_id=1, bos_token_id=0, ) return config, input_ids, batch_size def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case , _snake_case , _snake_case : Dict = self._get_config_and_data() _snake_case : List[str] = FlaxBlenderbotForConditionalGeneration(a_ ) _snake_case : str = lm_model(input_ids=a_ ) _snake_case : Optional[int] = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape, a_ ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = BlenderbotConfig( vocab_size=self.vocab_size, d_model=14, encoder_layers=2, decoder_layers=2, encoder_attention_heads=2, decoder_attention_heads=2, encoder_ffn_dim=8, decoder_ffn_dim=8, max_position_embeddings=48, ) _snake_case : Union[str, Any] = FlaxBlenderbotForConditionalGeneration(a_ ) _snake_case : str = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]], dtype=np.intaa ) _snake_case : Dict = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]], dtype=np.intaa ) _snake_case : Optional[Any] = lm_model(input_ids=a_, decoder_input_ids=a_ ) _snake_case : List[str] = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape, a_ ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : str = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]], dtype=np.intaa ) _snake_case : List[str] = shift_tokens_right(a_, 1, 2 ) _snake_case : int = np.equal(a_, 1 ).astype(np.floataa ).sum() _snake_case : Optional[int] = np.equal(a_, 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape, input_ids.shape ) self.assertEqual(a_, n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0], 2 ).all() ) @require_flax class lowercase( __a , unittest.TestCase , __a ): '''simple docstring''' lowercase__ = True lowercase__ = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowercase__ = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Union[str, Any] = FlaxBlenderbotModelTester(self ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case , _snake_case : List[str] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(a_, a_, a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case , _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(a_, a_, a_ ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case , _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case : List[Any] = self._prepare_for_class(a_, a_ ) _snake_case : List[Any] = model_class(a_ ) @jax.jit def encode_jitted(a_: str, a_: Optional[Any]=None, **a_: List[str] ): return model.encode(input_ids=a_, attention_mask=a_ ) with self.subTest("""JIT Enabled""" ): _snake_case : Optional[Any] = encode_jitted(**a_ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case : int = encode_jitted(**a_ ).to_tuple() self.assertEqual(len(a_ ), len(a_ ) ) for jitted_output, output in zip(a_, a_ ): self.assertEqual(jitted_output.shape, output.shape ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case , _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case : int = model_class(a_ ) _snake_case : int = model.encode(inputs_dict["""input_ids"""], inputs_dict["""attention_mask"""] ) _snake_case : Optional[Any] = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(a_: Any, a_: Tuple, a_: int ): return model.decode( decoder_input_ids=a_, decoder_attention_mask=a_, encoder_outputs=a_, ) with self.subTest("""JIT Enabled""" ): _snake_case : Any = decode_jitted(**a_ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case : List[str] = decode_jitted(**a_ ).to_tuple() self.assertEqual(len(a_ ), len(a_ ) ) for jitted_output, output in zip(a_, a_ ): self.assertEqual(jitted_output.shape, output.shape ) @slow def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _snake_case : List[Any] = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case : Optional[int] = np.ones((1, 1) ) * model.config.eos_token_id _snake_case : Union[str, Any] = model(a_ ) self.assertIsNotNone(a_ ) @unittest.skipUnless(jax_device != """cpu""", """3B test too slow on CPU.""" ) @slow def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : Dict = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _snake_case : str = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _snake_case : List[str] = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""", from_pt=a_ ) _snake_case : Union[str, Any] = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _snake_case : str = ["""Sam"""] _snake_case : Tuple = tokenizer(a_, return_tensors="""jax""" ) _snake_case : str = model.generate(**a_, **a_ ) _snake_case : List[Any] = """Sam is a great name. It means \"sun\" in Gaelic.""" _snake_case : int = tokenizer.batch_decode(a_, **a_ ) assert generated_txt[0].strip() == tgt_text
64
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str]=False ): """simple docstring""" _snake_case : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Dict , snake_case__ : List[str]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : List[Any] = """""" else: _snake_case : List[Any] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _snake_case : Union[str, Any] = in_proj_bias[: config.hidden_size] _snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : List[str] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" _snake_case : List[str] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : int ): """simple docstring""" _snake_case : Optional[Any] = dct.pop(snake_case__ ) _snake_case : Union[str, Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : str = ViTMSNConfig() _snake_case : Any = 10_00 _snake_case : Tuple = """datasets/huggingface/label-files""" _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : int = json.load(open(hf_hub_download(snake_case__ , snake_case__ ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[Any] = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _snake_case : Tuple = 3_84 _snake_case : Dict = 15_36 _snake_case : Tuple = 6 elif "l16" in checkpoint_url: _snake_case : Any = 10_24 _snake_case : int = 40_96 _snake_case : str = 24 _snake_case : Optional[int] = 16 _snake_case : List[Any] = 0.1 elif "b4" in checkpoint_url: _snake_case : Tuple = 4 elif "l7" in checkpoint_url: _snake_case : int = 7 _snake_case : Dict = 10_24 _snake_case : Optional[Any] = 40_96 _snake_case : Any = 24 _snake_case : Union[str, Any] = 16 _snake_case : Optional[int] = 0.1 _snake_case : int = ViTMSNModel(snake_case__ ) _snake_case : Optional[int] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" )["""target_encoder"""] _snake_case : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(snake_case__ ) _snake_case : List[str] = create_rename_keys(snake_case__ , base_model=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) read_in_q_k_v(snake_case__ , snake_case__ , base_model=snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Tuple = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) _snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=snake_case__ , image_std=snake_case__ ) _snake_case : Any = image_processor(images=snake_case__ , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) _snake_case : int = model(**snake_case__ ) _snake_case : List[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _snake_case : Optional[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: _snake_case : str = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: _snake_case : Optional[int] = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: _snake_case : List[Any] = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: _snake_case : Optional[int] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , snake_case__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
64
1
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" _snake_case : List[str] = 0 _snake_case : Optional[int] = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _snake_case : int = i + 1 else: _snake_case : List[str] = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(F'''{two_pointer([2, 7, 11, 15], 9) = }''')
64
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
1
"""simple docstring""" from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''google/efficientnet-b7''': '''https://huggingface.co/google/efficientnet-b7/resolve/main/config.json''', } class lowercase( __a ): '''simple docstring''' lowercase__ = "efficientnet" def __init__( self: Union[str, Any], a_: int = 3, a_: int = 600, a_: float = 2.0, a_: float = 3.1, a_: int = 8, a_: List[int] = [3, 3, 5, 3, 5, 5, 3], a_: List[int] = [32, 16, 24, 40, 80, 112, 192], a_: List[int] = [16, 24, 40, 80, 112, 192, 320], a_: List[int] = [], a_: List[int] = [1, 2, 2, 2, 1, 2, 1], a_: List[int] = [1, 2, 2, 3, 3, 4, 1], a_: List[int] = [1, 6, 6, 6, 6, 6, 6], a_: float = 0.25, a_: str = "swish", a_: int = 2_560, a_: str = "mean", a_: float = 0.02, a_: float = 0.001, a_: float = 0.99, a_: float = 0.5, a_: float = 0.2, **a_: Optional[int], ): '''simple docstring''' super().__init__(**a_ ) _snake_case : int = num_channels _snake_case : str = image_size _snake_case : List[str] = width_coefficient _snake_case : str = depth_coefficient _snake_case : Tuple = depth_divisor _snake_case : Optional[Any] = kernel_sizes _snake_case : int = in_channels _snake_case : List[str] = out_channels _snake_case : Optional[Any] = depthwise_padding _snake_case : List[Any] = strides _snake_case : str = num_block_repeats _snake_case : List[Any] = expand_ratios _snake_case : Tuple = squeeze_expansion_ratio _snake_case : Union[str, Any] = hidden_act _snake_case : Optional[int] = hidden_dim _snake_case : Dict = pooling_type _snake_case : Dict = initializer_range _snake_case : Dict = batch_norm_eps _snake_case : Optional[Any] = batch_norm_momentum _snake_case : str = dropout_rate _snake_case : Optional[int] = drop_connect_rate _snake_case : int = sum(a_ ) * 4 class lowercase( __a ): '''simple docstring''' lowercase__ = version.parse("1.11" ) @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return 1E-5
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" stooge(snake_case__ , 0 , len(snake_case__ ) - 1 ) return arr def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any , snake_case__ : int ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) # Recursively sort last 2/3 elements stooge(snake_case__ , i + t , (snake_case__) ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
64
1
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowercase( metaclass=__a ): '''simple docstring''' lowercase__ = ["note_seq"] def __init__( self: Dict, *a_: Union[str, Any], **a_: List[str] ): '''simple docstring''' requires_backends(self, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Optional[int], *a_: Any, **a_: Optional[Any] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Tuple, *a_: Optional[Any], **a_: List[str] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] )
64
1
"""simple docstring""" # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( '''stable diffusion controlnet''', '''0.22.0''', '''Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.''', standard_warn=False, stacklevel=3, )
64
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: List[str] ): '''simple docstring''' _snake_case : int = data _snake_case : Dict = [0X67452301, 0Xefcdab89, 0X98badcfe, 0X10325476, 0Xc3d2e1f0] @staticmethod def UpperCamelCase_ ( a_: Optional[Any], a_: Dict ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0Xffffffff def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _snake_case : Optional[int] = self.data + padding + struct.pack(""">Q""", 8 * len(self.data ) ) return padded_data def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0, len(self.padded_data ), 64 ) ] def UpperCamelCase_ ( self: Optional[Any], a_: List[Any] ): '''simple docstring''' _snake_case : List[str] = list(struct.unpack(""">16L""", a_ ) ) + [0] * 64 for i in range(16, 80 ): _snake_case : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]), 1 ) return w def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = self.padding() _snake_case : str = self.split_blocks() for block in self.blocks: _snake_case : Any = self.expand_block(a_ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = self.h for i in range(0, 80 ): if 0 <= i < 20: _snake_case : int = (b & c) | ((~b) & d) _snake_case : str = 0X5a827999 elif 20 <= i < 40: _snake_case : Optional[int] = b ^ c ^ d _snake_case : str = 0X6ed9eba1 elif 40 <= i < 60: _snake_case : List[Any] = (b & c) | (b & d) | (c & d) _snake_case : List[Any] = 0X8f1bbcdc elif 60 <= i < 80: _snake_case : List[Any] = b ^ c ^ d _snake_case : int = 0Xca62c1d6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = ( self.rotate(a_, 5 ) + f + e + k + expanded_block[i] & 0Xffffffff, a, self.rotate(a_, 30 ), c, d, ) _snake_case : Union[str, Any] = ( self.h[0] + a & 0Xffffffff, self.h[1] + b & 0Xffffffff, self.h[2] + c & 0Xffffffff, self.h[3] + d & 0Xffffffff, self.h[4] + e & 0Xffffffff, ) return ("{:08x}" * 5).format(*self.h ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = B"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) _snake_case : Union[str, Any] = parser.parse_args() _snake_case : List[Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: _snake_case : str = f.read() else: _snake_case : int = bytes(snake_case__ , """utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
64
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss A_ = pytest.mark.integration @require_faiss class lowercase( __a ): '''simple docstring''' def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : str = Dataset.from_dict({"""filename""": ["""my_name-train""" + """_""" + str(a_ ) for x in np.arange(30 ).tolist()]} ) return dset def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' import faiss _snake_case : Dataset = self._create_dummy_dataset() _snake_case : Any = dset.map( lambda a_, a_ : {"vecs": i * np.ones(5, dtype=np.floataa )}, with_indices=a_, keep_in_memory=a_ ) _snake_case : List[Any] = dset.add_faiss_index("""vecs""", batch_size=100, metric_type=faiss.METRIC_INNER_PRODUCT ) _snake_case , _snake_case : Optional[Any] = dset.get_nearest_examples("""vecs""", np.ones(5, dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0], """my_name-train_29""" ) dset.drop_index("""vecs""" ) def UpperCamelCase_ ( self: str ): '''simple docstring''' import faiss _snake_case : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1, 1 ), index_name="""vecs""", batch_size=100, metric_type=faiss.METRIC_INNER_PRODUCT, ) _snake_case , _snake_case : Any = dset.get_nearest_examples("""vecs""", np.ones(5, dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0], """my_name-train_29""" ) def UpperCamelCase_ ( self: int ): '''simple docstring''' import faiss _snake_case : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1, 1 ), index_name="""vecs""", metric_type=faiss.METRIC_INNER_PRODUCT, ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=a_ ) as tmp_file: dset.save_faiss_index("""vecs""", tmp_file.name ) dset.load_faiss_index("""vecs2""", tmp_file.name ) os.unlink(tmp_file.name ) _snake_case , _snake_case : List[str] = dset.get_nearest_examples("""vecs2""", np.ones(5, dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0], """my_name-train_29""" ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1, 1 ), index_name="""vecs""" ) dset.drop_index("""vecs""" ) self.assertRaises(a_, partial(dset.get_nearest_examples, """vecs2""", np.ones(5, dtype=np.floataa ) ) ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' from elasticsearch import Elasticsearch _snake_case : Dataset = self._create_dummy_dataset() with patch("""elasticsearch.Elasticsearch.search""" ) as mocked_search, patch( """elasticsearch.client.IndicesClient.create""" ) as mocked_index_create, patch("""elasticsearch.helpers.streaming_bulk""" ) as mocked_bulk: _snake_case : Tuple = {"""acknowledged""": True} mocked_bulk.return_value([(True, None)] * 30 ) _snake_case : List[str] = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 29}]}} _snake_case : Dict = Elasticsearch() dset.add_elasticsearch_index("""filename""", es_client=a_ ) _snake_case , _snake_case : int = dset.get_nearest_examples("""filename""", """my_name-train_29""" ) self.assertEqual(examples["""filename"""][0], """my_name-train_29""" ) @require_faiss class lowercase( __a ): '''simple docstring''' def UpperCamelCase_ ( self: str ): '''simple docstring''' import faiss _snake_case : int = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5, dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal, 5 ) index.add_vectors(np.zeros((5, 5), dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal, 10 ) # single query _snake_case : List[Any] = np.zeros(5, dtype=np.floataa ) _snake_case : List[str] = 1 _snake_case , _snake_case : str = index.search(a_ ) self.assertRaises(a_, index.search, query.reshape(-1, 1 ) ) self.assertGreater(scores[0], 0 ) self.assertEqual(indices[0], 1 ) # batched queries _snake_case : Union[str, Any] = np.eye(5, dtype=np.floataa )[::-1] _snake_case , _snake_case : Optional[Any] = index.search_batch(a_ ) self.assertRaises(a_, index.search_batch, queries[0] ) _snake_case : Optional[Any] = [scores[0] for scores in total_scores] _snake_case : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(a_ ), 0 ) self.assertListEqual([4, 3, 2, 1, 0], a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' import faiss _snake_case : List[str] = FaissIndex(string_factory="""Flat""" ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index, faiss.IndexFlat ) _snake_case : Any = FaissIndex(string_factory="""LSH""" ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index, faiss.IndexLSH ) with self.assertRaises(a_ ): _snake_case : Dict = FaissIndex(string_factory="""Flat""", custom_index=faiss.IndexFlat(5 ) ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' import faiss _snake_case : Any = faiss.IndexFlat(5 ) _snake_case : Tuple = FaissIndex(custom_index=a_ ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index, faiss.IndexFlat ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' import faiss _snake_case : List[str] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=a_ ) as tmp_file: index.save(tmp_file.name ) _snake_case : str = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) _snake_case : int = np.zeros(5, dtype=np.floataa ) _snake_case : List[str] = 1 _snake_case , _snake_case : List[Any] = index.search(a_ ) self.assertGreater(scores[0], 0 ) self.assertEqual(indices[0], 1 ) @require_faiss def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" import faiss _snake_case : Dict = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) _snake_case : Any = """index.faiss""" _snake_case : int = F"mock://{index_name}" index.save(snake_case__ , storage_options=mockfs.storage_options ) _snake_case : Optional[Any] = FaissIndex.load(snake_case__ , storage_options=mockfs.storage_options ) _snake_case : Union[str, Any] = np.zeros(5 , dtype=np.floataa ) _snake_case : List[str] = 1 _snake_case , _snake_case : str = index.search(snake_case__ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class lowercase( __a ): '''simple docstring''' def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' from elasticsearch import Elasticsearch with patch("""elasticsearch.Elasticsearch.search""" ) as mocked_search, patch( """elasticsearch.client.IndicesClient.create""" ) as mocked_index_create, patch("""elasticsearch.helpers.streaming_bulk""" ) as mocked_bulk: _snake_case : Tuple = Elasticsearch() _snake_case : Union[str, Any] = {"""acknowledged""": True} _snake_case : List[str] = ElasticSearchIndex(es_client=a_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["""foo""", """bar""", """foobar"""] ) # single query _snake_case : str = """foo""" _snake_case : Dict = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 0}]}} _snake_case , _snake_case : Tuple = index.search(a_ ) self.assertEqual(scores[0], 1 ) self.assertEqual(indices[0], 0 ) # single query with timeout _snake_case : Union[str, Any] = """foo""" _snake_case : Tuple = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 0}]}} _snake_case , _snake_case : int = index.search(a_, request_timeout=30 ) self.assertEqual(scores[0], 1 ) self.assertEqual(indices[0], 0 ) # batched queries _snake_case : str = ["""foo""", """bar""", """foobar"""] _snake_case : Dict = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 1}]}} _snake_case , _snake_case : Tuple = index.search_batch(a_ ) _snake_case : List[Any] = [scores[0] for scores in total_scores] _snake_case : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(a_ ), 0 ) self.assertListEqual([1, 1, 1], a_ ) # batched queries with timeout _snake_case : Optional[int] = ["""foo""", """bar""", """foobar"""] _snake_case : List[Any] = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 1}]}} _snake_case , _snake_case : Tuple = index.search_batch(a_, request_timeout=30 ) _snake_case : Union[str, Any] = [scores[0] for scores in total_scores] _snake_case : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(a_ ), 0 ) self.assertListEqual([1, 1, 1], a_ )
64
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor A_ = transforms.Compose( [ transforms.Resize((2_56, 2_56)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if isinstance(snake_case__ , torch.Tensor ): return image elif isinstance(snake_case__ , PIL.Image.Image ): _snake_case : Tuple = [image] _snake_case : str = [trans(img.convert("""RGB""" ) ) for img in image] _snake_case : Dict = torch.stack(snake_case__ ) return image class lowercase( __a ): '''simple docstring''' def __init__( self: Tuple, a_: Optional[Any], a_: int ): '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _snake_case : Union[str, Any] = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=a_, scheduler=a_ ) def UpperCamelCase_ ( self: Tuple, a_: Tuple ): '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f"The value of strength should in [0.0, 1.0] but is {strength}" ) def UpperCamelCase_ ( self: str, a_: Union[str, Any], a_: Optional[Any], a_: int ): '''simple docstring''' _snake_case : str = min(int(num_inference_steps * strength ), a_ ) _snake_case : Optional[Any] = max(num_inference_steps - init_timestep, 0 ) _snake_case : str = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCamelCase_ ( self: Union[str, Any], a_: str, a_: List[Any], a_: str, a_: Optional[int], a_: str, a_: Optional[Any]=None ): '''simple docstring''' if not isinstance(a_, (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f"`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(a_ )}" ) _snake_case : Any = image.to(device=a_, dtype=a_ ) if isinstance(a_, a_ ) and len(a_ ) != batch_size: raise ValueError( f"You have passed a list of generators of length {len(a_ )}, but requested an effective batch" f" size of {batch_size}. Make sure the batch size matches the length of the generators." ) _snake_case : Any = init_latents.shape _snake_case : List[str] = randn_tensor(a_, generator=a_, device=a_, dtype=a_ ) # get latents print("""add noise to latents at timestep""", a_ ) _snake_case : Dict = self.scheduler.add_noise(a_, a_, a_ ) _snake_case : str = init_latents return latents @torch.no_grad() def __call__( self: Optional[int], a_: Union[torch.FloatTensor, PIL.Image.Image] = None, a_: float = 0.8, a_: int = 1, a_: Optional[Union[torch.Generator, List[torch.Generator]]] = None, a_: float = 0.0, a_: int = 50, a_: Optional[bool] = None, a_: Optional[str] = "pil", a_: bool = True, ): '''simple docstring''' self.check_inputs(a_ ) # 2. Preprocess image _snake_case : Dict = preprocess(a_ ) # 3. set timesteps self.scheduler.set_timesteps(a_, device=self.device ) _snake_case , _snake_case : Optional[Any] = self.get_timesteps(a_, a_, self.device ) _snake_case : Optional[int] = timesteps[:1].repeat(a_ ) # 4. Prepare latent variables _snake_case : Dict = self.prepare_latents(a_, a_, a_, self.unet.dtype, self.device, a_ ) _snake_case : Optional[int] = latents # 5. Denoising loop for t in self.progress_bar(a_ ): # 1. predict noise model_output _snake_case : Dict = self.unet(a_, a_ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _snake_case : Dict = self.scheduler.step( a_, a_, a_, eta=a_, use_clipped_model_output=a_, generator=a_, ).prev_sample _snake_case : Dict = (image / 2 + 0.5).clamp(0, 1 ) _snake_case : Union[str, Any] = image.cpu().permute(0, 2, 3, 1 ).numpy() if output_type == "pil": _snake_case : Any = self.numpy_to_pil(a_ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=a_ )
64
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
1
"""simple docstring""" import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowercase( __a ): '''simple docstring''' lowercase__ = (DDIMParallelScheduler,) lowercase__ = (("eta", 0.0), ("num_inference_steps", 50)) def UpperCamelCase_ ( self: str, **a_: int ): '''simple docstring''' _snake_case : int = { """num_train_timesteps""": 1_000, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", """clip_sample""": True, } config.update(**a_ ) return config def UpperCamelCase_ ( self: int, **a_: Tuple ): '''simple docstring''' _snake_case : List[str] = self.scheduler_classes[0] _snake_case : List[str] = self.get_scheduler_config(**a_ ) _snake_case : List[Any] = scheduler_class(**a_ ) _snake_case , _snake_case : Union[str, Any] = 10, 0.0 _snake_case : List[Any] = self.dummy_model() _snake_case : Tuple = self.dummy_sample_deter scheduler.set_timesteps(a_ ) for t in scheduler.timesteps: _snake_case : Optional[Any] = model(a_, a_ ) _snake_case : Optional[int] = scheduler.step(a_, a_, a_, a_ ).prev_sample return sample def UpperCamelCase_ ( self: int ): '''simple docstring''' for timesteps in [100, 500, 1_000]: self.check_over_configs(num_train_timesteps=a_ ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=a_ ) _snake_case : List[Any] = self.scheduler_classes[0] _snake_case : str = self.get_scheduler_config(steps_offset=1 ) _snake_case : List[Any] = scheduler_class(**a_ ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps, torch.LongTensor([801, 601, 401, 201, 1] ) ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1], [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=a_, beta_end=a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a_ ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a_ ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=a_ ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' self.check_over_configs(thresholding=a_ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=a_, prediction_type=a_, sample_max_value=a_, ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' for t in [1, 10, 49]: self.check_over_forward(time_step=a_ ) def UpperCamelCase_ ( self: int ): '''simple docstring''' for t, num_inference_steps in zip([1, 10, 50], [10, 50, 500] ): self.check_over_forward(time_step=a_, num_inference_steps=a_ ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' for t, eta in zip([1, 10, 49], [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=a_, eta=a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Any = self.scheduler_classes[0] _snake_case : List[str] = self.get_scheduler_config() _snake_case : Tuple = scheduler_class(**a_ ) assert torch.sum(torch.abs(scheduler._get_variance(0, 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420, 400 ) - 0.14_771 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980, 960 ) - 0.32_460 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0, 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487, 486 ) - 0.00_979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999, 998 ) - 0.02 ) ) < 1E-5 def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Optional[Any] = self.scheduler_classes[0] _snake_case : Any = self.get_scheduler_config() _snake_case : List[Any] = scheduler_class(**a_ ) _snake_case , _snake_case : List[Any] = 10, 0.0 scheduler.set_timesteps(a_ ) _snake_case : Optional[Any] = self.dummy_model() _snake_case : Union[str, Any] = self.dummy_sample_deter _snake_case : Dict = self.dummy_sample_deter + 0.1 _snake_case : Union[str, Any] = self.dummy_sample_deter - 0.1 _snake_case : Any = samplea.shape[0] _snake_case : Optional[Any] = torch.stack([samplea, samplea, samplea], dim=0 ) _snake_case : Dict = torch.arange(a_ )[0:3, None].repeat(1, a_ ) _snake_case : List[Any] = model(samples.flatten(0, 1 ), timesteps.flatten(0, 1 ) ) _snake_case : List[str] = scheduler.batch_step_no_noise(a_, timesteps.flatten(0, 1 ), samples.flatten(0, 1 ), a_ ) _snake_case : Tuple = torch.sum(torch.abs(a_ ) ) _snake_case : str = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 1_147.7_904 ) < 1E-2 assert abs(result_mean.item() - 0.4_982 ) < 1E-3 def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Dict = self.full_loop() _snake_case : Union[str, Any] = torch.sum(torch.abs(a_ ) ) _snake_case : int = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 172.0_067 ) < 1E-2 assert abs(result_mean.item() - 0.223_967 ) < 1E-3 def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : List[Any] = self.full_loop(prediction_type="""v_prediction""" ) _snake_case : Optional[int] = torch.sum(torch.abs(a_ ) ) _snake_case : List[Any] = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 52.5_302 ) < 1E-2 assert abs(result_mean.item() - 0.0_684 ) < 1E-3 def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = self.full_loop(set_alpha_to_one=a_, beta_start=0.01 ) _snake_case : Dict = torch.sum(torch.abs(a_ ) ) _snake_case : List[str] = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 149.8_295 ) < 1E-2 assert abs(result_mean.item() - 0.1_951 ) < 1E-3 def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Optional[int] = self.full_loop(set_alpha_to_one=a_, beta_start=0.01 ) _snake_case : List[Any] = torch.sum(torch.abs(a_ ) ) _snake_case : Optional[int] = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 149.0_784 ) < 1E-2 assert abs(result_mean.item() - 0.1_941 ) < 1E-3
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list ): """simple docstring""" if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] _snake_case : List[Any] = [] def generate(snake_case__ : int , snake_case__ : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _snake_case , _snake_case : Optional[Any] = arr[k - 1], arr[i] else: # k is odd _snake_case , _snake_case : List[str] = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
64
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''facebook/data2vec-vision-base-ft''': ( '''https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json''' ), } class lowercase( __a ): '''simple docstring''' lowercase__ = "data2vec-vision" def __init__( self: Optional[Any], a_: Tuple=768, a_: int=12, a_: Any=12, a_: Union[str, Any]=3_072, a_: Union[str, Any]="gelu", a_: Union[str, Any]=0.0, a_: str=0.0, a_: int=0.02, a_: Optional[int]=1E-12, a_: Dict=224, a_: Optional[int]=16, a_: Any=3, a_: List[Any]=False, a_: int=False, a_: Dict=False, a_: Union[str, Any]=False, a_: List[str]=0.1, a_: int=0.1, a_: Optional[int]=True, a_: List[Any]=[3, 5, 7, 11], a_: List[Any]=[1, 2, 3, 6], a_: int=True, a_: int=0.4, a_: int=256, a_: Optional[Any]=1, a_: Tuple=False, a_: str=255, **a_: Dict, ): '''simple docstring''' super().__init__(**a_ ) _snake_case : Optional[int] = hidden_size _snake_case : List[Any] = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : Any = intermediate_size _snake_case : int = hidden_act _snake_case : Optional[Any] = hidden_dropout_prob _snake_case : List[str] = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : Tuple = layer_norm_eps _snake_case : Tuple = image_size _snake_case : Optional[int] = patch_size _snake_case : Optional[Any] = num_channels _snake_case : Any = use_mask_token _snake_case : str = use_absolute_position_embeddings _snake_case : Optional[int] = use_relative_position_bias _snake_case : List[str] = use_shared_relative_position_bias _snake_case : Tuple = layer_scale_init_value _snake_case : Optional[Any] = drop_path_rate _snake_case : Union[str, Any] = use_mean_pooling # decode head attributes (semantic segmentation) _snake_case : List[Any] = out_indices _snake_case : List[str] = pool_scales # auxiliary head attributes (semantic segmentation) _snake_case : Tuple = use_auxiliary_head _snake_case : List[str] = auxiliary_loss_weight _snake_case : List[str] = auxiliary_channels _snake_case : List[Any] = auxiliary_num_convs _snake_case : List[str] = auxiliary_concat_input _snake_case : List[str] = semantic_loss_ignore_index class lowercase( __a ): '''simple docstring''' lowercase__ = version.parse("1.11" ) @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return 1E-4
64
"""simple docstring""" from math import factorial A_ = {str(d): factorial(d) for d in range(10)} def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" return sum(DIGIT_FACTORIAL[d] for d in str(snake_case__ ) ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , snake_case__ ) if sum_of_digit_factorial(snake_case__ ) == i ) if __name__ == "__main__": print(F'''{solution() = }''')
64
1
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Optional[Any]=False ): """simple docstring""" _snake_case : Any = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("""head""" ): _snake_case : List[str] = """segformer.encoder.""" + key if key.startswith("""backbone""" ): _snake_case : int = key.replace("""backbone""" , """segformer.encoder""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _snake_case : Optional[int] = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] _snake_case : Dict = key.replace(F"patch_embed{idx}" , F"patch_embeddings.{int(snake_case__ )-1}" ) if "norm" in key: _snake_case : Union[str, Any] = key.replace("""norm""" , """layer_norm""" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _snake_case : str = key[key.find("""segformer.encoder.layer_norm""" ) + len("""segformer.encoder.layer_norm""" )] _snake_case : int = key.replace(F"layer_norm{idx}" , F"layer_norm.{int(snake_case__ )-1}" ) if "layer_norm1" in key: _snake_case : List[Any] = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: _snake_case : int = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 _snake_case : str = key[key.find("""block""" ) + len("""block""" )] _snake_case : str = key.replace(F"block{idx}" , F"block.{int(snake_case__ )-1}" ) if "attn.q" in key: _snake_case : Any = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: _snake_case : int = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: _snake_case : int = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: _snake_case : Optional[Any] = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: _snake_case : Optional[Any] = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: _snake_case : Optional[Any] = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: _snake_case : Union[str, Any] = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) _snake_case : List[Any] = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _snake_case : int = key[key.find("""linear_c""" ) + len("""linear_c""" )] _snake_case : Optional[Any] = key.replace(F"linear_c{idx}" , F"linear_c.{int(snake_case__ )-1}" ) if key.startswith("""head""" ): _snake_case : Optional[int] = key.replace("""head""" , """classifier""" ) _snake_case : Optional[Any] = value return new_state_dict def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Dict ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _snake_case : Tuple = state_dict.pop(F"segformer.encoder.block.{i}.{j}.attention.self.kv.weight" ) _snake_case : Tuple = state_dict.pop(F"segformer.encoder.block.{i}.{j}.attention.self.kv.bias" ) # next, add keys and values (in that order) to the state dict _snake_case : List[Any] = kv_weight[ : config.hidden_sizes[i], : ] _snake_case : List[str] = kv_bias[: config.hidden_sizes[i]] _snake_case : Any = kv_weight[ config.hidden_sizes[i] :, : ] _snake_case : Optional[Any] = kv_bias[ config.hidden_sizes[i] : ] def UpperCAmelCase__ (): """simple docstring""" _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : str = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return image @torch.no_grad() def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : int , snake_case__ : List[str] ): """simple docstring""" _snake_case : Tuple = SegformerConfig() _snake_case : Optional[Any] = False # set attributes based on model_name _snake_case : Optional[Any] = """huggingface/label-files""" if "segformer" in model_name: _snake_case : Tuple = model_name[len("""segformer.""" ) : len("""segformer.""" ) + 2] if "ade" in model_name: _snake_case : Optional[Any] = 1_50 _snake_case : Tuple = """ade20k-id2label.json""" _snake_case : List[Any] = (1, 1_50, 1_28, 1_28) elif "city" in model_name: _snake_case : Optional[int] = 19 _snake_case : Dict = """cityscapes-id2label.json""" _snake_case : str = (1, 19, 1_28, 1_28) else: raise ValueError(F"Model {model_name} not supported" ) elif "mit" in model_name: _snake_case : Union[str, Any] = True _snake_case : List[str] = model_name[4:6] _snake_case : Dict = 10_00 _snake_case : Optional[int] = """imagenet-1k-id2label.json""" _snake_case : Tuple = (1, 10_00) else: raise ValueError(F"Model {model_name} not supported" ) # set config attributes _snake_case : Optional[int] = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : int = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _snake_case : Optional[int] = [64, 1_28, 3_20, 5_12] _snake_case : Optional[int] = 2_56 elif size == "b2": _snake_case : Any = [64, 1_28, 3_20, 5_12] _snake_case : List[str] = 7_68 _snake_case : List[str] = [3, 4, 6, 3] elif size == "b3": _snake_case : List[Any] = [64, 1_28, 3_20, 5_12] _snake_case : str = 7_68 _snake_case : List[Any] = [3, 4, 18, 3] elif size == "b4": _snake_case : Any = [64, 1_28, 3_20, 5_12] _snake_case : Any = 7_68 _snake_case : Dict = [3, 8, 27, 3] elif size == "b5": _snake_case : int = [64, 1_28, 3_20, 5_12] _snake_case : str = 7_68 _snake_case : List[str] = [3, 6, 40, 3] else: raise ValueError(F"Size {size} not supported" ) # load image processor (only resize + normalize) _snake_case : Optional[int] = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=snake_case__ , align=snake_case__ , do_random_crop=snake_case__ ) # prepare image _snake_case : Dict = prepare_img() _snake_case : List[Any] = image_processor(images=snake_case__ , return_tensors="""pt""" ).pixel_values logger.info(F"Converting model {model_name}..." ) # load original state dict if encoder_only: _snake_case : Dict = torch.load(snake_case__ , map_location=torch.device("""cpu""" ) ) else: _snake_case : str = torch.load(snake_case__ , map_location=torch.device("""cpu""" ) )["""state_dict"""] # rename keys _snake_case : int = rename_keys(snake_case__ , encoder_only=snake_case__ ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(snake_case__ , snake_case__ ) # create HuggingFace model and load state dict if encoder_only: _snake_case : List[Any] = False _snake_case : Optional[Any] = SegformerForImageClassification(snake_case__ ) else: _snake_case : List[str] = SegformerForSemanticSegmentation(snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() # forward pass _snake_case : List[Any] = model(snake_case__ ) _snake_case : Any = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _snake_case : Tuple = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _snake_case : Union[str, Any] = torch.tensor( [ [[-7.58_20, -8.72_31, -8.32_15], [-8.06_00, -10.35_29, -10.03_04], [-7.52_08, -9.41_03, -9.62_39]], [[-12.69_18, -13.89_94, -13.71_37], [-13.31_96, -15.75_23, -15.47_89], [-12.93_43, -14.87_57, -14.96_89]], [[-11.19_11, -11.94_21, -11.32_43], [-11.33_42, -13.68_39, -13.35_81], [-10.39_09, -12.18_32, -12.48_58]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _snake_case : List[Any] = torch.tensor( [ [[-11.81_73, -14.38_50, -16.31_28], [-14.56_48, -16.58_04, -18.65_68], [-14.72_23, -15.73_87, -18.42_18]], [[-15.72_90, -17.91_71, -19.44_23], [-18.31_05, -19.94_48, -21.46_61], [-17.92_96, -18.64_97, -20.79_10]], [[-15.07_83, -17.03_36, -18.27_89], [-16.87_71, -18.68_70, -20.16_12], [-16.24_54, -17.14_26, -19.50_55]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _snake_case : Union[str, Any] = torch.tensor( [ [[-9.08_78, -10.20_81, -10.18_91], [-9.31_44, -10.79_41, -10.98_43], [-9.22_94, -10.38_55, -10.57_04]], [[-12.23_16, -13.90_68, -13.61_02], [-12.91_61, -14.37_02, -14.32_35], [-12.52_33, -13.71_74, -13.79_32]], [[-14.62_75, -15.24_90, -14.97_27], [-14.34_00, -15.96_87, -16.28_27], [-14.14_84, -15.40_33, -15.89_37]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _snake_case : Union[str, Any] = torch.tensor( [ [[-12.31_44, -13.24_47, -14.08_02], [-13.36_14, -14.58_16, -15.61_17], [-13.33_40, -14.44_33, -16.22_19]], [[-19.27_81, -20.41_28, -20.75_06], [-20.61_53, -21.65_66, -22.09_98], [-19.98_00, -21.04_30, -22.14_94]], [[-18.87_39, -19.78_04, -21.18_34], [-20.12_33, -21.67_65, -23.29_44], [-20.03_15, -21.26_41, -23.69_44]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _snake_case : List[Any] = torch.tensor( [ [[-9.55_24, -12.08_35, -11.73_48], [-10.52_29, -13.64_46, -14.56_62], [-9.58_42, -12.88_51, -13.94_14]], [[-15.34_32, -17.53_23, -17.08_18], [-16.33_30, -18.92_55, -19.21_01], [-15.13_40, -17.78_48, -18.39_71]], [[-12.60_72, -14.94_86, -14.66_31], [-13.76_29, -17.09_07, -17.77_45], [-12.78_99, -16.16_95, -17.16_71]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _snake_case : Optional[Any] = torch.tensor( [ [[-11.92_95, -13.40_57, -14.81_06], [-13.34_31, -14.81_79, -15.37_81], [-14.28_36, -15.59_42, -16.15_88]], [[-11.49_06, -12.80_67, -13.65_64], [-13.11_89, -14.05_00, -14.15_43], [-13.87_48, -14.51_36, -14.87_89]], [[0.53_74, 0.10_67, -0.47_42], [0.11_41, -0.22_55, -0.70_99], [-0.30_00, -0.59_24, -1.31_05]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _snake_case : Optional[Any] = torch.tensor( [ [[-7.82_17, -9.87_67, -10.17_17], [-9.44_38, -10.90_58, -11.40_47], [-9.79_39, -12.34_95, -12.10_79]], [[-7.15_14, -9.53_36, -10.08_60], [-9.77_76, -11.68_22, -11.84_39], [-10.14_11, -12.76_55, -12.89_72]], [[0.30_21, 0.08_05, -0.23_10], [-0.03_28, -0.16_05, -0.27_14], [-0.14_08, -0.54_77, -0.69_76]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _snake_case : str = torch.tensor( [ [ [-1.1_3_7_2e0_1, -1.2_7_8_7e0_1, -1.3_4_7_7e0_1], [-1.2_5_3_6e0_1, -1.4_1_9_4e0_1, -1.4_4_0_9e0_1], [-1.3_2_1_7e0_1, -1.4_8_8_8e0_1, -1.5_3_2_7e0_1], ], [ [-1.4_7_9_1e0_1, -1.7_1_2_2e0_1, -1.8_2_7_7e0_1], [-1.7_1_6_3e0_1, -1.9_1_9_2e0_1, -1.9_5_3_3e0_1], [-1.7_8_9_7e0_1, -1.9_9_9_1e0_1, -2.0_3_1_5e0_1], ], [ [7.6_7_2_3e-0_1, 4.1_9_2_1e-0_1, -7.7_8_7_8e-0_2], [4.7_7_7_2e-0_1, 9.5_5_5_7e-0_3, -2.8_0_8_2e-0_1], [3.6_0_3_2e-0_1, -2.4_8_2_6e-0_1, -5.1_1_6_8e-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _snake_case : Any = torch.tensor( [ [[-9.49_59, -11.30_87, -11.74_79], [-11.00_25, -12.65_40, -12.33_19], [-11.40_64, -13.04_87, -12.99_05]], [[-9.89_05, -11.30_84, -12.08_54], [-11.17_26, -12.76_98, -12.95_83], [-11.59_85, -13.32_78, -14.17_74]], [[0.22_13, 0.01_92, -0.24_66], [-0.17_31, -0.42_13, -0.48_74], [-0.31_26, -0.65_41, -1.13_89]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _snake_case : Tuple = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _snake_case : List[Any] = torch.tensor( [ [[-16.09_76, -16.48_56, -17.39_62], [-16.62_34, -19.03_42, -19.76_85], [-16.09_00, -18.06_61, -19.11_80]], [[-18.47_50, -18.84_88, -19.50_74], [-19.40_30, -22.15_70, -22.59_77], [-19.11_91, -20.84_86, -22.37_83]], [[-4.51_78, -5.50_37, -6.51_09], [-5.08_84, -7.21_74, -8.03_34], [-4.41_56, -5.81_17, -7.29_70]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _snake_case : int = torch.tensor( [ [[-14.20_81, -14.47_32, -14.19_77], [-14.58_67, -16.44_23, -16.63_56], [-13.44_41, -14.96_85, -16.86_96]], [[-14.45_76, -14.70_73, -15.04_51], [-15.08_16, -17.62_37, -17.98_73], [-14.42_13, -16.01_99, -18.59_92]], [[-4.73_49, -4.95_88, -5.09_66], [-4.32_10, -6.93_25, -7.25_91], [-3.43_12, -4.74_84, -7.19_17]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _snake_case : Tuple = torch.tensor( [ [[-11.77_37, -11.95_26, -11.32_73], [-13.66_92, -14.45_74, -13.88_78], [-13.89_37, -14.69_24, -15.93_45]], [[-14.67_06, -14.53_30, -14.13_06], [-16.15_02, -16.81_80, -16.42_69], [-16.83_38, -17.89_39, -20.17_46]], [[1.04_91, 0.82_89, 1.03_10], [1.10_44, 0.52_19, 0.80_55], [1.08_99, 0.69_26, 0.55_90]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _snake_case : List[Any] = torch.tensor( [ [[-12.56_41, -13.47_77, -13.06_84], [-13.95_87, -15.89_83, -16.65_57], [-13.31_09, -15.73_50, -16.31_41]], [[-14.70_74, -15.43_52, -14.59_44], [-16.63_53, -18.16_63, -18.61_20], [-15.17_02, -18.03_29, -18.15_47]], [[-1.79_90, -2.09_51, -1.77_84], [-2.63_97, -3.82_45, -3.96_86], [-1.52_64, -2.81_26, -2.93_16]], ] ) else: _snake_case : Optional[int] = logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1e-2 ) # finally, save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) A_ = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" if len(snake_case__ ) < k or k < 0: raise ValueError("""Invalid Input""" ) _snake_case : Optional[int] = sum(array[:k] ) for i in range(len(snake_case__ ) - k ): _snake_case : Optional[Any] = current_sum - array[i] + array[i + k] _snake_case : List[str] = max(snake_case__ , snake_case__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-10_00, 10_00) for i in range(1_00)] A_ = randint(0, 1_10) print(F'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
64
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ = {'''configuration_timm_backbone''': ['''TimmBackboneConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''TimmBackbone'''] if TYPE_CHECKING: from .configuration_timm_backbone import TimmBackboneConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timm_backbone import TimmBackbone else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/vocab.json''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/vocab.json''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/vocab.json''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/vocab.json''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/vocab.json''', }, '''merges_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/merges.txt''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/merges.txt''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/merges.txt''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/merges.txt''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/tokenizer.json''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/tokenizer.json''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/tokenizer.json''', }, } A_ = { '''gpt2''': 10_24, '''gpt2-medium''': 10_24, '''gpt2-large''': 10_24, '''gpt2-xl''': 10_24, '''distilgpt2''': 10_24, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = GPTaTokenizer def __init__( self: Any, a_: Optional[Any]=None, a_: Tuple=None, a_: Optional[int]=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: Optional[Any]="<|endoftext|>", a_: List[str]=False, **a_: Any, ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) _snake_case : List[Any] = kwargs.pop("""add_bos_token""", a_ ) _snake_case : Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Any = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Union[str, Any] = add_prefix_space _snake_case : Optional[int] = pre_tok_class(**a_ ) _snake_case : Any = add_prefix_space def UpperCamelCase_ ( self: Tuple, *a_: List[str], **a_: Tuple ): '''simple docstring''' _snake_case : Optional[Any] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Optional[Any], **a_: Dict ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: List[Any], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Any = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: int, a_: "Conversation" ): '''simple docstring''' _snake_case : Any = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(a_, add_special_tokens=a_ ) + [self.eos_token_id] ) if len(a_ ) > self.model_max_length: _snake_case : List[Any] = input_ids[-self.model_max_length :] return input_ids
64
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = CodeGenTokenizer lowercase__ = CodeGenTokenizerFast lowercase__ = True lowercase__ = {"add_prefix_space": True} lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : Tuple = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] _snake_case : Tuple = dict(zip(a_, range(len(a_ ) ) ) ) _snake_case : str = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _snake_case : List[Any] = {"""unk_token""": """<unk>"""} _snake_case : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(a_ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(a_ ) ) def UpperCamelCase_ ( self: Any, **a_: int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Any, **a_: str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Dict ): '''simple docstring''' _snake_case : Union[str, Any] = """lower newer""" _snake_case : Tuple = """lower newer""" return input_text, output_text def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = CodeGenTokenizer(self.vocab_file, self.merges_file, **self.special_tokens_map ) _snake_case : Optional[Any] = """lower newer""" _snake_case : Optional[int] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _snake_case : int = tokenizer.tokenize(a_, add_prefix_space=a_ ) self.assertListEqual(a_, a_ ) _snake_case : str = tokens + [tokenizer.unk_token] _snake_case : Optional[int] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : int = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : Dict = """lower newer""" # Testing tokenization _snake_case : Dict = tokenizer.tokenize(a_, add_prefix_space=a_ ) _snake_case : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids without special tokens _snake_case : Optional[Any] = tokenizer.encode(a_, add_special_tokens=a_, add_prefix_space=a_ ) _snake_case : Tuple = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids with special tokens _snake_case : Tuple = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : int = tokenizer.encode(a_, add_prefix_space=a_ ) _snake_case : Optional[Any] = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) # Testing the unknown token _snake_case : Tuple = tokens + [rust_tokenizer.unk_token] _snake_case : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: List[Any]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) # Simple input _snake_case : Any = """This is a simple input""" _snake_case : Optional[int] = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") _snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) # Pair input self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname, pad_token="""<pad>""" ) # Simple input _snake_case : List[Any] = """This is a simple input""" _snake_case : int = ["""This is a simple input looooooooong""", """This is a simple input"""] _snake_case : Any = ("""This is a simple input""", """This is a pair""") _snake_case : str = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] _snake_case : str = tokenizer.pad_token_id _snake_case : Optional[int] = tokenizer(a_, padding="""max_length""", max_length=30, return_tensors="""np""" ) _snake_case : Dict = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) _snake_case : Tuple = tokenizer(*a_, padding="""max_length""", max_length=60, return_tensors="""np""" ) _snake_case : Optional[Any] = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1], 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1], 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1], 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1], 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = """$$$""" _snake_case : List[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname, bos_token=a_, add_bos_token=a_ ) _snake_case : str = """This is a simple input""" _snake_case : int = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Tuple = tokenizer(a_ ) _snake_case : Optional[Any] = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0], a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _snake_case : Optional[int] = tokenizer.decode(out_s.input_ids ) _snake_case : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0], a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) _snake_case : Dict = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" _snake_case : Union[str, Any] = """\nif len_a > len_b: result = a\nelse: result = b""" _snake_case : Optional[Any] = tokenizer.encode(a_ ) _snake_case : Dict = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] _snake_case : Optional[Any] = tokenizer.decode(a_, truncate_before_pattern=a_ ) self.assertEqual(a_, a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass
64
1
"""simple docstring""" import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = 0 lowercase__ = False lowercase__ = 3.0 class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' self.assertDictEqual(MockClass().to_kwargs(), {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs(), {"""a""": 2} ) self.assertDictEqual(MockClass(a=2, b=a_ ).to_kwargs(), {"""a""": 2, """b""": True} ) self.assertDictEqual(MockClass(a=2, c=2.25 ).to_kwargs(), {"""a""": 2, """c""": 2.25} ) @require_cuda def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : int = GradScalerKwargs(init_scale=1_024, growth_factor=2 ) AcceleratorState._reset_state() _snake_case : str = Accelerator(mixed_precision="""fp16""", kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) _snake_case : Union[str, Any] = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale, 1_024.0 ) self.assertEqual(scaler._growth_factor, 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor, 0.5 ) self.assertEqual(scaler._growth_interval, 2_000 ) self.assertEqual(scaler._enabled, a_ ) @require_multi_gpu def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Optional[Any] = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] execute_subprocess_async(a_, env=os.environ.copy() ) if __name__ == "__main__": A_ = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) A_ = Accelerator(kwargs_handlers=[ddp_scaler]) A_ = torch.nn.Linear(1_00, 2_00) A_ = accelerator.prepare(model) # Check the values changed in kwargs A_ = '''''' A_ = model.bucket_bytes_cap // (10_24 * 10_24) if observed_bucket_cap_map != 15: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
64
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ = re.compile(r'''\s+''') def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" return {"hash": hashlib.mda(re.sub(snake_case__ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : Any = [len(snake_case__ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(snake_case__ ), "line_max": max(snake_case__ )} def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : List[Any] ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : List[str]=5 ): """simple docstring""" _snake_case : Any = ["""auto-generated""", """autogenerated""", """automatically generated"""] _snake_case : Tuple = example["""content"""].splitlines() for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Union[str, Any]=5 , snake_case__ : Any=0.05 ): """simple docstring""" _snake_case : Optional[Any] = ["""unit tests""", """test file""", """configuration file"""] _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : Dict = 0 _snake_case : str = 0 # first test for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _snake_case : Optional[int] = example["""content"""].count("""\n""" ) _snake_case : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[int] = ["""def """, """class """, """for """, """while """] _snake_case : str = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : List[str]=4 ): """simple docstring""" _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : str = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" _snake_case : Optional[Any] = tokenizer(example["""content"""] , truncation=snake_case__ )["""input_ids"""] _snake_case : Optional[Any] = len(example["""content"""] ) / len(snake_case__ ) return {"ratio": ratio} def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = {} results.update(get_hash(snake_case__ ) ) results.update(line_stats(snake_case__ ) ) results.update(alpha_stats(snake_case__ ) ) results.update(char_token_ratio(snake_case__ ) ) results.update(is_autogenerated(snake_case__ ) ) results.update(is_config_or_test(snake_case__ ) ) results.update(has_no_keywords(snake_case__ ) ) results.update(has_few_assignments(snake_case__ ) ) return results def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] ): """simple docstring""" if not check_uniques(snake_case__ , snake_case__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """rb""" ) as f_in: with gzip.open(str(snake_case__ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(snake_case__ , snake_case__ ) os.unlink(snake_case__ ) # Settings A_ = HfArgumentParser(PreprocessingArguments) A_ = parser.parse_args() if args.num_workers is None: A_ = multiprocessing.cpu_count() A_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ = time.time() A_ = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ = time.time() A_ = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ = set(ds.unique('''hash''')) A_ = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ = time.time() A_ = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ = time.time() A_ , A_ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) A_ = output_dir / '''data''' data_dir.mkdir(exist_ok=True) A_ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ = str(data_dir / F'''file-{file_number+1:012}.json''') A_ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
64
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = 42 lowercase__ = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.26.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version('''>=''', '''0.0.12''') ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = 42 lowercase__ = 42 from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
64
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
1
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" if len(snake_case__ ) < k or k < 0: raise ValueError("""Invalid Input""" ) _snake_case : Optional[int] = sum(array[:k] ) for i in range(len(snake_case__ ) - k ): _snake_case : Optional[Any] = current_sum - array[i] + array[i + k] _snake_case : List[str] = max(snake_case__ , snake_case__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-10_00, 10_00) for i in range(1_00)] A_ = randint(0, 1_10) print(F'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
64
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : str , snake_case__ : str ): """simple docstring""" for attribute in key.split(""".""" ): _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ).shape else: _snake_case : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": _snake_case : int = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : Tuple = value elif weight_type == "bias": _snake_case : List[str] = value else: _snake_case : int = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str] ): """simple docstring""" _snake_case : List[Any] = [] _snake_case : Optional[Any] = fairseq_model.state_dict() _snake_case : str = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _snake_case : Optional[Any] = None for name, value in fairseq_dict.items(): _snake_case : Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) _snake_case : Dict = True elif name.split(""".""" )[0] == "proj": _snake_case : Dict = fairseq_model.proj _snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: _snake_case : Dict = True if "*" in mapped_key: _snake_case : Optional[int] = name.split(snake_case__ )[0].split(""".""" )[-2] _snake_case : Union[str, Any] = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: _snake_case : str = """weight_g""" elif "weight_v" in name: _snake_case : Optional[Any] = """weight_v""" elif "bias" in name: _snake_case : Union[str, Any] = """bias""" elif "weight" in name: _snake_case : int = """weight""" else: _snake_case : Optional[int] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Dict , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] , snake_case__ : int ): """simple docstring""" _snake_case : Any = full_name.split("""conv_layers.""" )[-1] _snake_case : Optional[int] = name.split(""".""" ) _snake_case : List[str] = int(items[0] ) _snake_case : Dict = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _snake_case : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _snake_case : int = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _snake_case : List[str] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" _snake_case , _snake_case : Optional[Any] = emb.weight.shape _snake_case : Optional[int] = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) _snake_case : Union[str, Any] = emb.weight.data return lin_layer def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Any = f.readlines() _snake_case : Optional[Any] = [line.split(""" """ )[0] for line in lines] _snake_case : str = len(snake_case__ ) _snake_case : Tuple = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Union[str, Any] , ): """simple docstring""" _snake_case : Optional[int] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : List[str] = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) _snake_case : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) _snake_case , _snake_case , _snake_case : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) _snake_case : Optional[Any] = model[0].eval() # set weights for wav2vec2 encoder _snake_case : Any = WavaVecaModel(snake_case__ ) _snake_case : Optional[Any] = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) _snake_case : Optional[Any] = SpeechaTextaForCausalLM(snake_case__ ) _snake_case , _snake_case : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) _snake_case : Any = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) _snake_case : Any = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) _snake_case : Any = False # add projection layer _snake_case : int = nn.Parameter(projection_layer.weight ) _snake_case : Any = nn.Parameter(projection_layer.bias ) _snake_case : Any = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) _snake_case : Dict = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) _snake_case : str = hf_wavavec.config.to_dict() _snake_case : List[str] = tokenizer.pad_token_id _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Union[str, Any] = tokenizer.eos_token_id _snake_case : Optional[Any] = """speech_to_text_2""" _snake_case : Optional[int] = """wav2vec2""" _snake_case : Tuple = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_02_24, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') A_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
64
1
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Tuple = inspect.getfile(accelerate.test_utils ) _snake_case : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) _snake_case : Optional[Any] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) _snake_case : int = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices." ) _snake_case : Optional[int] = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: Dict ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices." ) _snake_case : Optional[int] = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.operation_file_path] print(f"Command: {cmd}" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : int = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_, env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' print(f"Found {torch.cuda.device_count()} devices, using 2 devices only" ) _snake_case : Tuple = ["""torchrun""", f"--nproc_per_node={torch.cuda.device_count()}", self.data_loop_file_path] with patch_environment(omp_num_threads=1, cuda_visible_devices="""0,1""" ): execute_subprocess_async(a_, env=os.environ.copy() ) if __name__ == "__main__": A_ = Accelerator() A_ = (accelerator.state.process_index + 2, 10) A_ = torch.randint(0, 10, shape).to(accelerator.device) A_ = '''''' A_ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." A_ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." A_ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
64
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = [ ('''bert.bert''', '''visual_bert'''), ('''bert.cls''', '''cls'''), ('''bert.classifier''', '''cls'''), ('''token_type_embeddings_visual''', '''visual_token_type_embeddings'''), ('''position_embeddings_visual''', '''visual_position_embeddings'''), ('''projection''', '''visual_projection'''), ] A_ = [ '''nlvr2_coco_pre_trained.th''', '''nlvr2_fine_tuned.th''', '''nlvr2_pre_trained.th''', '''vcr_coco_pre_train.th''', '''vcr_fine_tune.th''', '''vcr_pre_train.th''', '''vqa_coco_pre_trained.th''', '''vqa_fine_tuned.th''', '''vqa_pre_trained.th''', ] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : str = torch.load(snake_case__ , map_location="""cpu""" ) return sd def UpperCAmelCase__ (snake_case__ : int , snake_case__ : str , snake_case__ : int=rename_keys_prefix ): """simple docstring""" _snake_case : Optional[int] = OrderedDict() _snake_case : List[str] = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _snake_case : Any = key for name_pair in rename_keys_prefix: _snake_case : str = new_key.replace(name_pair[0] , name_pair[1] ) _snake_case : Union[str, Any] = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _snake_case : List[Any] = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : str ): """simple docstring""" assert ( checkpoint_path.split("""/""" )[-1] in ACCEPTABLE_CHECKPOINTS ), F"The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}." # Get Config if "pre" in checkpoint_path: _snake_case : Union[str, Any] = """pretraining""" if "vcr" in checkpoint_path: _snake_case : int = {"""visual_embedding_dim""": 5_12} elif "vqa_advanced" in checkpoint_path: _snake_case : Optional[int] = {"""visual_embedding_dim""": 20_48} elif "vqa" in checkpoint_path: _snake_case : Any = {"""visual_embedding_dim""": 20_48} elif "nlvr" in checkpoint_path: _snake_case : List[Any] = {"""visual_embedding_dim""": 10_24} else: raise NotImplementedError(F"No implementation found for `{checkpoint_path}`." ) else: if "vcr" in checkpoint_path: _snake_case : Dict = {"""visual_embedding_dim""": 5_12} _snake_case : Union[str, Any] = """multichoice""" elif "vqa_advanced" in checkpoint_path: _snake_case : Dict = {"""visual_embedding_dim""": 20_48} _snake_case : Optional[Any] = """vqa_advanced""" elif "vqa" in checkpoint_path: _snake_case : Optional[Any] = {"""visual_embedding_dim""": 20_48, """num_labels""": 31_29} _snake_case : List[str] = """vqa""" elif "nlvr" in checkpoint_path: _snake_case : List[Any] = { """visual_embedding_dim""": 10_24, """num_labels""": 2, } _snake_case : List[Any] = """nlvr""" _snake_case : Optional[int] = VisualBertConfig(**snake_case__ ) # Load State Dict _snake_case : List[str] = load_state_dict(snake_case__ ) _snake_case : Any = get_new_dict(snake_case__ , snake_case__ ) if model_type == "pretraining": _snake_case : Union[str, Any] = VisualBertForPreTraining(snake_case__ ) elif model_type == "vqa": _snake_case : Any = VisualBertForQuestionAnswering(snake_case__ ) elif model_type == "nlvr": _snake_case : Optional[Any] = VisualBertForVisualReasoning(snake_case__ ) elif model_type == "multichoice": _snake_case : Optional[int] = VisualBertForMultipleChoice(snake_case__ ) model.load_state_dict(snake_case__ ) # Save Checkpoints Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument('''orig_checkpoint_path''', type=str, help='''A path to .th on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', type=str, help='''Path to the output PyTorch model.''') A_ = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
64
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
1
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowercase( __a ): '''simple docstring''' def __init__( self: Tuple, a_: Union[str, Any], a_: Tuple, a_: Optional[Any]=1_024, a_: Union[str, Any]=1_024, a_: Any=3.6 ): '''simple docstring''' _snake_case : List[str] = tokenizer _snake_case : str = tokenizer.bos_token_id _snake_case : List[str] = dataset _snake_case : Union[str, Any] = seq_length _snake_case : List[Any] = seq_length * chars_per_token * num_of_sequences def __iter__( self: Dict ): '''simple docstring''' _snake_case : Any = iter(self.dataset ) _snake_case : Tuple = True while more_examples: _snake_case , _snake_case : List[Any] = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(a_ )["""content"""] ) buffer_len += len(buffer[-1] ) except StopIteration: _snake_case : List[Any] = False break _snake_case : Any = tokenizer(a_, truncation=a_ )["""input_ids"""] _snake_case : Optional[Any] = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0, len(a_ ), self.seq_length ): _snake_case : int = all_token_ids[i : i + self.seq_length] if len(a_ ) == self.seq_length: yield torch.tensor(a_ ) def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Dict = {"""streaming""": True} _snake_case : Dict = load_dataset(args.dataset_name , split="""train""" , **snake_case__ ) _snake_case : Optional[int] = ConstantLengthDataset(snake_case__ , snake_case__ , seq_length=args.seq_length ) _snake_case : Optional[Any] = DataLoader(snake_case__ , batch_size=args.batch_size ) return eval_dataloader def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" model.eval() _snake_case : Any = [] for step, batch in enumerate(snake_case__ ): with torch.no_grad(): _snake_case : Tuple = model(snake_case__ , labels=snake_case__ ) _snake_case : List[Any] = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(snake_case__ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _snake_case : Tuple = torch.mean(torch.cat(snake_case__ ) ) try: _snake_case : Union[str, Any] = torch.exp(snake_case__ ) except OverflowError: _snake_case : List[Any] = float("""inf""" ) return loss.item(), perplexity.item() # Setup Accelerator A_ = Accelerator() # Parse configuration A_ = HfArgumentParser(EvaluationArguments) A_ = parser.parse_args() set_seed(args.seed) # Logging A_ = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer A_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) A_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader A_ = create_dataloader(args) # Prepare everything with our `accelerator`. A_ , A_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') A_ , A_ = evaluate(args) logger.info(F'''loss/eval: {eval_loss}, perplexity: {perplexity}''')
64
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
1
"""simple docstring""" import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput A_ = '''scheduler_config.json''' class lowercase( __a ): '''simple docstring''' lowercase__ = 1 lowercase__ = 2 lowercase__ = 3 lowercase__ = 4 lowercase__ = 5 lowercase__ = 6 lowercase__ = 7 lowercase__ = 8 lowercase__ = 9 lowercase__ = 10 lowercase__ = 11 lowercase__ = 12 lowercase__ = 13 lowercase__ = 14 @dataclass class lowercase( __a ): '''simple docstring''' lowercase__ = 42 class lowercase: '''simple docstring''' lowercase__ = SCHEDULER_CONFIG_NAME lowercase__ = [] lowercase__ = True @classmethod def UpperCamelCase_ ( cls: Optional[int], a_: Dict[str, Any] = None, a_: Optional[str] = None, a_: Any=False, **a_: Optional[Any], ): '''simple docstring''' _snake_case , _snake_case , _snake_case : Tuple = cls.load_config( pretrained_model_name_or_path=a_, subfolder=a_, return_unused_kwargs=a_, return_commit_hash=a_, **a_, ) return cls.from_config(a_, return_unused_kwargs=a_, **a_ ) def UpperCamelCase_ ( self: List[Any], a_: Union[str, os.PathLike], a_: bool = False, **a_: Union[str, Any] ): '''simple docstring''' self.save_config(save_directory=a_, push_to_hub=a_, **a_ ) @property def UpperCamelCase_ ( self: str ): '''simple docstring''' return self._get_compatibles() @classmethod def UpperCamelCase_ ( cls: List[str] ): '''simple docstring''' _snake_case : Any = list(set([cls.__name__] + cls._compatibles ) ) _snake_case : Union[str, Any] = importlib.import_module(__name__.split(""".""" )[0] ) _snake_case : Optional[Any] = [ getattr(a_, a_ ) for c in compatible_classes_str if hasattr(a_, a_ ) ] return compatible_classes
64
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
1
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): A_ = '''pt''' elif is_tf_available(): A_ = '''tf''' else: A_ = '''jax''' class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = PerceiverTokenizer lowercase__ = False def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' super().setUp() _snake_case : Any = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return PerceiverTokenizer.from_pretrained("""deepmind/language-perceiver""" ) def UpperCamelCase_ ( self: List[str], **a_: Any ): '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Dict, a_: Any, a_: Tuple=False, a_: List[str]=20, a_: Tuple=5 ): '''simple docstring''' _snake_case : Any = [] for i in range(len(a_ ) ): try: _snake_case : Any = tokenizer.decode([i], clean_up_tokenization_spaces=a_ ) except UnicodeDecodeError: pass toks.append((i, tok) ) _snake_case : int = list(filter(lambda a_ : re.match(r"""^[ a-zA-Z]+$""", t[1] ), a_ ) ) _snake_case : Optional[Any] = list(filter(lambda a_ : [t[0]] == tokenizer.encode(t[1], add_special_tokens=a_ ), a_ ) ) if max_length is not None and len(a_ ) > max_length: _snake_case : List[str] = toks[:max_length] if min_length is not None and len(a_ ) < min_length and len(a_ ) > 0: while len(a_ ) < min_length: _snake_case : Optional[Any] = toks + toks # toks_str = [t[1] for t in toks] _snake_case : List[Any] = [t[0] for t in toks] # Ensure consistency _snake_case : Tuple = tokenizer.decode(a_, clean_up_tokenization_spaces=a_ ) if " " not in output_txt and len(a_ ) > 1: _snake_case : Any = ( tokenizer.decode([toks_ids[0]], clean_up_tokenization_spaces=a_ ) + """ """ + tokenizer.decode(toks_ids[1:], clean_up_tokenization_spaces=a_ ) ) if with_prefix_space: _snake_case : str = """ """ + output_txt _snake_case : Dict = tokenizer.encode(a_, add_special_tokens=a_ ) return output_txt, output_ids def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Tuple = self.perceiver_tokenizer _snake_case : Union[str, Any] = """Unicode €.""" _snake_case : Dict = tokenizer(a_ ) _snake_case : List[str] = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded["""input_ids"""], a_ ) # decoding _snake_case : List[str] = tokenizer.decode(a_ ) self.assertEqual(a_, """[CLS]Unicode €.[SEP]""" ) _snake_case : Any = tokenizer("""e è é ê ë""" ) _snake_case : Union[str, Any] = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded["""input_ids"""], a_ ) # decoding _snake_case : Tuple = tokenizer.decode(a_ ) self.assertEqual(a_, """[CLS]e è é ê ë[SEP]""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ), """[CLS]e è é ê ë[SEP]""" ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Tuple = self.perceiver_tokenizer _snake_case : Any = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off _snake_case : Any = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on _snake_case : Tuple = tokenizer(a_, padding=a_, return_tensors=a_ ) self.assertIsInstance(a_, a_ ) if FRAMEWORK != "jax": _snake_case : str = list(batch.input_ids.numpy()[0] ) else: _snake_case : Any = list(batch.input_ids.tolist()[0] ) self.assertListEqual(a_, a_ ) self.assertEqual((2, 38), batch.input_ids.shape ) self.assertEqual((2, 38), batch.attention_mask.shape ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : int = self.perceiver_tokenizer _snake_case : str = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] _snake_case : Union[str, Any] = tokenizer(a_, padding=a_, return_tensors=a_ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""", a_ ) self.assertIn("""attention_mask""", a_ ) self.assertNotIn("""decoder_input_ids""", a_ ) self.assertNotIn("""decoder_attention_mask""", a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Dict = self.perceiver_tokenizer _snake_case : int = [ """Summary of the text.""", """Another summary.""", ] _snake_case : List[Any] = tokenizer( text_target=a_, max_length=32, padding="""max_length""", truncation=a_, return_tensors=a_ ) self.assertEqual(32, targets["""input_ids"""].shape[1] ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): self.assertNotEqual(tokenizer.model_max_length, 42 ) # Now let's start the test _snake_case : int = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc _snake_case : str = tempfile.mkdtemp() _snake_case : Tuple = """ He is very happy, UNwant\u00E9d,running""" _snake_case : Any = tokenizer.encode(a_, add_special_tokens=a_ ) tokenizer.save_pretrained(a_ ) _snake_case : List[str] = tokenizer.__class__.from_pretrained(a_ ) _snake_case : Tuple = after_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) shutil.rmtree(a_ ) _snake_case : Optional[int] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc _snake_case : Tuple = tempfile.mkdtemp() _snake_case : Optional[int] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) _snake_case : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) _snake_case : Tuple = tokenizer.encode(a_, add_special_tokens=a_ ) tokenizer.save_pretrained(a_ ) _snake_case : int = tokenizer.__class__.from_pretrained(a_ ) _snake_case : int = after_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) self.assertIn("""new_additional_special_token""", after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length, 42 ) _snake_case : int = tokenizer.__class__.from_pretrained(a_, model_max_length=43 ) self.assertEqual(tokenizer.model_max_length, 43 ) shutil.rmtree(a_ ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : int = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(a_ ) with open(os.path.join(a_, """special_tokens_map.json""" ), encoding="""utf-8""" ) as json_file: _snake_case : str = json.load(a_ ) with open(os.path.join(a_, """tokenizer_config.json""" ), encoding="""utf-8""" ) as json_file: _snake_case : Optional[Any] = json.load(a_ ) _snake_case : List[Any] = [f"<extra_id_{i}>" for i in range(125 )] _snake_case : int = added_tokens_extra_ids + [ """an_additional_special_token""" ] _snake_case : Optional[int] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(a_, """special_tokens_map.json""" ), """w""", encoding="""utf-8""" ) as outfile: json.dump(a_, a_ ) with open(os.path.join(a_, """tokenizer_config.json""" ), """w""", encoding="""utf-8""" ) as outfile: json.dump(a_, a_ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files _snake_case : Tuple = tokenizer_class.from_pretrained( a_, ) self.assertIn( """an_additional_special_token""", tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ["""an_additional_special_token"""], tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ), ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained _snake_case : str = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""", lstrip=a_ )] _snake_case : Any = tokenizer_class.from_pretrained( a_, additional_special_tokens=a_, ) self.assertIn("""a_new_additional_special_token""", tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""], tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ), ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : List[str] = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ), """�""" ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' pass def UpperCamelCase_ ( self: str ): '''simple docstring''' pass def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Dict = self.get_tokenizers(fast=a_, do_lower_case=a_ ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): _snake_case : Tuple = ["""[CLS]""", """t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """s""", """t""", """[SEP]"""] _snake_case : List[str] = tokenizer.convert_tokens_to_string(a_ ) self.assertIsInstance(a_, a_ )
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
1
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin A_ = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A_ = 25_00_04 A_ = 25_00_20 @require_sentencepiece @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = MBartaaTokenizer lowercase__ = MBartaaTokenizerFast lowercase__ = True lowercase__ = True def UpperCamelCase_ ( self: str ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _snake_case : List[str] = MBartaaTokenizer(a_, src_lang="""en_XX""", tgt_lang="""ro_RO""", keep_accents=a_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : Any = """<s>""" _snake_case : List[str] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ), a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ), a_ ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], """<s>""" ) self.assertEqual(vocab_keys[1], """<pad>""" ) self.assertEqual(vocab_keys[-1], """<mask>""" ) self.assertEqual(len(a_ ), 1_054 ) def UpperCamelCase_ ( self: str ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size, 1_054 ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : List[Any] = MBartaaTokenizer(a_, src_lang="""en_XX""", tgt_lang="""ro_RO""", keep_accents=a_ ) _snake_case : List[Any] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a_, ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a_ ), [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]], ) _snake_case : List[Any] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a_, [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """."""], ) _snake_case : Any = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_, [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ], ) _snake_case : Any = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_, [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """."""], ) @slow def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Union[str, Any] = {"""input_ids""": [[250_004, 11_062, 82_772, 7, 15, 82_772, 538, 51_529, 237, 17_198, 1_290, 206, 9, 215_175, 1_314, 136, 17_198, 1_290, 206, 9, 56_359, 42, 122_009, 9, 16_466, 16, 87_344, 4_537, 9, 4_717, 78_381, 6, 159_958, 7, 15, 24_480, 618, 4, 527, 22_693, 5_428, 4, 2_777, 24_480, 9_874, 4, 43_523, 594, 4, 803, 18_392, 33_189, 18, 4, 43_523, 24_447, 12_399, 100, 24_955, 83_658, 9_626, 144_057, 15, 839, 22_335, 16, 136, 24_955, 83_658, 83_479, 15, 39_102, 724, 16, 678, 645, 2_789, 1_328, 4_589, 42, 122_009, 115_774, 23, 805, 1_328, 46_876, 7, 136, 53_894, 1_940, 42_227, 41_159, 17_721, 823, 425, 4, 27_512, 98_722, 206, 136, 5_531, 4_970, 919, 17_336, 5, 2], [250_004, 20_080, 618, 83, 82_775, 47, 479, 9, 1_517, 73, 53_894, 333, 80_581, 110_117, 18_811, 5_256, 1_295, 51, 152_526, 297, 7_986, 390, 124_416, 538, 35_431, 214, 98, 15_044, 25_737, 136, 7_108, 43_701, 23, 756, 135_355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [250_004, 581, 63_773, 119_455, 6, 147_797, 88_203, 7, 645, 70, 21, 3_285, 10_269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_, model_name="""facebook/mbart-large-50""", revision="""d3913889c59cd5c9e456b269c376325eabad57e2""", ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _snake_case : Dict = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-mbart50""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : Dict = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : List[str] = self.tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : List[str] = tempfile.mkdtemp() _snake_case : Tuple = tokenizer_r.save_pretrained(a_ ) _snake_case : Tuple = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) _snake_case : List[Any] = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(a_, a_ ) # Checks everything loads correctly in the same way _snake_case : List[Any] = tokenizer_r.from_pretrained(a_ ) _snake_case : Any = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=True _snake_case : Any = tempfile.mkdtemp() _snake_case : List[str] = tokenizer_r.save_pretrained(a_, legacy_format=a_ ) _snake_case : List[str] = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files self.assertSequenceEqual(a_, a_ ) # Checks everything loads correctly in the same way _snake_case : Any = tokenizer_r.from_pretrained(a_ ) _snake_case : List[Any] = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=False _snake_case : List[str] = tempfile.mkdtemp() _snake_case : Dict = tokenizer_r.save_pretrained(a_, legacy_format=a_ ) _snake_case : Optional[Any] = tokenizer_p.save_pretrained(a_ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _snake_case : Dict = tokenizer_r.from_pretrained(a_ ) _snake_case : int = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) shutil.rmtree(a_ ) @require_torch @require_sentencepiece @require_tokenizers class lowercase( unittest.TestCase ): '''simple docstring''' lowercase__ = "facebook/mbart-large-50-one-to-many-mmt" lowercase__ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase__ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase__ = [EN_CODE, 82_74, 12_78_73, 2_59_16, 7, 86_22, 20_71, 4_38, 6_74_85, 53, 18_78_95, 23, 5_17_12, 2] @classmethod def UpperCamelCase_ ( cls: Optional[Any] ): '''simple docstring''' _snake_case : MBartaaTokenizer = MBartaaTokenizer.from_pretrained( cls.checkpoint_name, src_lang="""en_XX""", tgt_lang="""ro_RO""" ) _snake_case : Union[str, Any] = 1 return cls def UpperCamelCase_ ( self: int ): '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ar_AR"""], 250_001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""en_EN"""], 250_004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ro_RO"""], 250_020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""mr_IN"""], 250_038 ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : str = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens, a_ ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' self.assertIn(a_, self.tokenizer.all_special_ids ) _snake_case : Dict = [RO_CODE, 884, 9_019, 96, 9, 916, 86_792, 36, 18_743, 15_596, 5, 2] _snake_case : Dict = self.tokenizer.decode(a_, skip_special_tokens=a_ ) _snake_case : Tuple = self.tokenizer.decode(generated_ids[1:], skip_special_tokens=a_ ) self.assertEqual(a_, a_ ) self.assertNotIn(self.tokenizer.eos_token, a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Tuple = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0], a_ ) _snake_case : Any = 10 _snake_case : List[Any] = self.tokenizer(a_, max_length=a_, truncation=a_ ).input_ids[0] self.assertEqual(ids[0], a_ ) self.assertEqual(ids[-1], 2 ) self.assertEqual(len(a_ ), a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ), [250_053, 250_001] ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : List[str] = tempfile.mkdtemp() _snake_case : Optional[int] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(a_ ) _snake_case : Any = MBartaaTokenizer.from_pretrained(a_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids, a_ ) @require_torch def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : int = self.tokenizer(self.src_text, text_target=self.tgt_text, padding=a_, return_tensors="""pt""" ) _snake_case : Any = shift_tokens_right(batch["""labels"""], self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = self.tokenizer( self.src_text, text_target=self.tgt_text, padding=a_, truncation=a_, max_length=len(self.expected_src_tokens ), return_tensors="""pt""", ) _snake_case : List[Any] = shift_tokens_right(batch["""labels"""], self.tokenizer.pad_token_id ) self.assertIsInstance(a_, a_ ) self.assertEqual((2, 14), batch.input_ids.shape ) self.assertEqual((2, 14), batch.attention_mask.shape ) _snake_case : Optional[Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens, a_ ) self.assertEqual(2, batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens, [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens, [self.tokenizer.eos_token_id] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : int = self.tokenizer(self.src_text, padding=a_, truncation=a_, max_length=3, return_tensors="""pt""" ) _snake_case : int = self.tokenizer( text_target=self.tgt_text, padding=a_, truncation=a_, max_length=10, return_tensors="""pt""" ) _snake_case : Union[str, Any] = targets["""input_ids"""] _snake_case : Optional[int] = shift_tokens_right(a_, self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1], 3 ) self.assertEqual(batch.decoder_input_ids.shape[1], 10 ) @require_torch def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Tuple = self.tokenizer._build_translation_inputs( """A test""", return_tensors="""pt""", src_lang="""en_XX""", tgt_lang="""ar_AR""" ) self.assertEqual( nested_simplify(a_ ), { # en_XX, A, test, EOS """input_ids""": [[250_004, 62, 3_034, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 250_001, }, )
64
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = CodeGenTokenizer def __init__( self: Union[str, Any], a_: List[Any]=None, a_: str=None, a_: str=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: str="<|endoftext|>", a_: List[Any]=False, **a_: List[str], ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) if kwargs.pop("""add_bos_token""", a_ ): _snake_case : str = kwargs.pop("""name_or_path""", """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" f"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" f"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) _snake_case : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Dict = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Dict = add_prefix_space _snake_case : str = pre_tok_class(**a_ ) _snake_case : List[Any] = add_prefix_space def UpperCamelCase_ ( self: Any, *a_: Any, **a_: int ): '''simple docstring''' _snake_case : Optional[int] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Any, **a_: List[str] ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: str, a_: Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"], a_: bool = False, a_: bool = None, a_: Optional[List[str]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = super().decode( token_ids=a_, skip_special_tokens=a_, clean_up_tokenization_spaces=a_, **a_, ) if truncate_before_pattern is not None and len(a_ ) > 0: _snake_case : List[str] = self.truncate(a_, a_ ) return decoded_text def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Optional[Any] ): '''simple docstring''' def find_re(a_: Dict, a_: str, a_: Union[str, Any] ): _snake_case : Any = pattern.search(a_, a_ ) return m.start() if m else -1 _snake_case : Tuple = [re.compile(a_, re.MULTILINE ) for pattern in truncate_before_pattern] _snake_case : List[Any] = list(re.finditer("""^print""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : int = completion[: prints[1].start()] _snake_case : List[str] = list(re.finditer("""^def""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : List[Any] = completion[: defs[1].start()] _snake_case : int = 0 _snake_case : List[Any] = [ pos for pos in [find_re(a_, a_, a_ ) for terminal in terminals] if pos != -1 ] if len(a_ ) > 0: return completion[: min(a_ )] else: return completion
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list ): """simple docstring""" if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] _snake_case : List[Any] = [] def generate(snake_case__ : int , snake_case__ : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _snake_case , _snake_case : Optional[Any] = arr[k - 1], arr[i] else: # k is odd _snake_case , _snake_case : List[str] = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
64
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A_ = { '''configuration_poolformer''': [ '''POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PoolFormerConfig''', '''PoolFormerOnnxConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''PoolFormerFeatureExtractor'''] A_ = ['''PoolFormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PoolFormerForImageClassification''', '''PoolFormerModel''', '''PoolFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
64
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str]=False ): """simple docstring""" _snake_case : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Dict , snake_case__ : List[str]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : List[Any] = """""" else: _snake_case : List[Any] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _snake_case : Union[str, Any] = in_proj_bias[: config.hidden_size] _snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : List[str] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" _snake_case : List[str] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : int ): """simple docstring""" _snake_case : Optional[Any] = dct.pop(snake_case__ ) _snake_case : Union[str, Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : str = ViTMSNConfig() _snake_case : Any = 10_00 _snake_case : Tuple = """datasets/huggingface/label-files""" _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : int = json.load(open(hf_hub_download(snake_case__ , snake_case__ ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[Any] = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _snake_case : Tuple = 3_84 _snake_case : Dict = 15_36 _snake_case : Tuple = 6 elif "l16" in checkpoint_url: _snake_case : Any = 10_24 _snake_case : int = 40_96 _snake_case : str = 24 _snake_case : Optional[int] = 16 _snake_case : List[Any] = 0.1 elif "b4" in checkpoint_url: _snake_case : Tuple = 4 elif "l7" in checkpoint_url: _snake_case : int = 7 _snake_case : Dict = 10_24 _snake_case : Optional[Any] = 40_96 _snake_case : Any = 24 _snake_case : Union[str, Any] = 16 _snake_case : Optional[int] = 0.1 _snake_case : int = ViTMSNModel(snake_case__ ) _snake_case : Optional[int] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" )["""target_encoder"""] _snake_case : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(snake_case__ ) _snake_case : List[str] = create_rename_keys(snake_case__ , base_model=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) read_in_q_k_v(snake_case__ , snake_case__ , base_model=snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Tuple = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) _snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=snake_case__ , image_std=snake_case__ ) _snake_case : Any = image_processor(images=snake_case__ , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) _snake_case : int = model(**snake_case__ ) _snake_case : List[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _snake_case : Optional[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: _snake_case : str = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: _snake_case : Optional[int] = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: _snake_case : List[Any] = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: _snake_case : Optional[int] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , snake_case__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
64
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''uclanlp/visualbert-vqa''': '''https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json''', '''uclanlp/visualbert-vqa-pre''': '''https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json''', '''uclanlp/visualbert-vqa-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-vcr''': '''https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json''', '''uclanlp/visualbert-vcr-pre''': '''https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json''', '''uclanlp/visualbert-vcr-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-nlvr2''': '''https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-pre''': '''https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json''' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class lowercase( __a ): '''simple docstring''' lowercase__ = "visual_bert" def __init__( self: Optional[int], a_: Optional[int]=30_522, a_: Tuple=768, a_: List[Any]=512, a_: List[Any]=12, a_: Dict=12, a_: Union[str, Any]=3_072, a_: List[str]="gelu", a_: Optional[Any]=0.1, a_: Optional[int]=0.1, a_: str=512, a_: int=2, a_: List[Any]=0.02, a_: Union[str, Any]=1E-12, a_: List[str]=False, a_: Dict=True, a_: Optional[Any]=1, a_: List[str]=0, a_: Any=2, **a_: List[Any], ): '''simple docstring''' super().__init__(pad_token_id=a_, bos_token_id=a_, eos_token_id=a_, **a_ ) _snake_case : Tuple = vocab_size _snake_case : Optional[Any] = max_position_embeddings _snake_case : Optional[int] = hidden_size _snake_case : int = visual_embedding_dim _snake_case : str = num_hidden_layers _snake_case : str = num_attention_heads _snake_case : str = intermediate_size _snake_case : Any = hidden_act _snake_case : int = hidden_dropout_prob _snake_case : str = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : str = type_vocab_size _snake_case : Optional[int] = layer_norm_eps _snake_case : Optional[int] = bypass_transformer _snake_case : Union[str, Any] = special_visual_initialize
64
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
1
"""simple docstring""" import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str="shi-labs/oneformer_demo" ): """simple docstring""" with open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) as f: _snake_case : Optional[int] = json.load(snake_case__ ) _snake_case : str = {} _snake_case : Union[str, Any] = [] _snake_case : str = [] for key, info in class_info.items(): _snake_case : str = info["""name"""] class_names.append(info["""name"""] ) if info["isthing"]: thing_ids.append(int(snake_case__ ) ) _snake_case : Optional[Any] = thing_ids _snake_case : str = class_names return metadata class lowercase( unittest.TestCase ): '''simple docstring''' def __init__( self: List[str], a_: Dict, a_: Any=7, a_: int=3, a_: List[Any]=30, a_: Tuple=400, a_: int=None, a_: List[Any]=True, a_: Dict=True, a_: Any=[0.5, 0.5, 0.5], a_: str=[0.5, 0.5, 0.5], a_: Dict=10, a_: List[str]=False, a_: Optional[Any]=255, a_: Optional[Any]="shi-labs/oneformer_demo", a_: Optional[Any]="ade20k_panoptic.json", a_: int=10, ): '''simple docstring''' _snake_case : List[Any] = parent _snake_case : Tuple = batch_size _snake_case : str = num_channels _snake_case : int = min_resolution _snake_case : Tuple = max_resolution _snake_case : int = do_resize _snake_case : Optional[Any] = {"""shortest_edge""": 32, """longest_edge""": 1_333} if size is None else size _snake_case : Union[str, Any] = do_normalize _snake_case : Union[str, Any] = image_mean _snake_case : Dict = image_std _snake_case : List[str] = class_info_file _snake_case : Dict = prepare_metadata(a_, a_ ) _snake_case : List[Any] = num_text _snake_case : Tuple = repo_path # for the post_process_functions _snake_case : Tuple = 2 _snake_case : Any = 10 _snake_case : Optional[int] = 10 _snake_case : Any = 3 _snake_case : str = 4 _snake_case : List[str] = num_labels _snake_case : Any = do_reduce_labels _snake_case : Union[str, Any] = ignore_index def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def UpperCamelCase_ ( self: int, a_: str, a_: Optional[int]=False ): '''simple docstring''' if not batched: _snake_case : Tuple = image_inputs[0] if isinstance(a_, Image.Image ): _snake_case , _snake_case : List[str] = image.size else: _snake_case , _snake_case : Optional[int] = image.shape[1], image.shape[2] if w < h: _snake_case : Union[str, Any] = int(self.size["""shortest_edge"""] * h / w ) _snake_case : Optional[Any] = self.size["""shortest_edge"""] elif w > h: _snake_case : List[str] = self.size["""shortest_edge"""] _snake_case : List[Any] = int(self.size["""shortest_edge"""] * w / h ) else: _snake_case : List[str] = self.size["""shortest_edge"""] _snake_case : List[Any] = self.size["""shortest_edge"""] else: _snake_case : Optional[int] = [] for image in image_inputs: _snake_case , _snake_case : Any = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _snake_case : Tuple = max(a_, key=lambda a_ : item[0] )[0] _snake_case : str = max(a_, key=lambda a_ : item[1] )[1] return expected_height, expected_width def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ), masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ), ) @require_torch @require_vision class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string lowercase__ = image_processing_class def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : List[str] = OneFormerImageProcessorTester(self ) @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' return self.image_processing_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a_, """image_mean""" ) ) self.assertTrue(hasattr(a_, """image_std""" ) ) self.assertTrue(hasattr(a_, """do_normalize""" ) ) self.assertTrue(hasattr(a_, """do_resize""" ) ) self.assertTrue(hasattr(a_, """size""" ) ) self.assertTrue(hasattr(a_, """ignore_index""" ) ) self.assertTrue(hasattr(a_, """class_info_file""" ) ) self.assertTrue(hasattr(a_, """num_text""" ) ) self.assertTrue(hasattr(a_, """repo_path""" ) ) self.assertTrue(hasattr(a_, """metadata""" ) ) self.assertTrue(hasattr(a_, """do_reduce_labels""" ) ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _snake_case : str = prepare_image_inputs(self.image_processing_tester, equal_resolution=a_ ) for image in image_inputs: self.assertIsInstance(a_, Image.Image ) # Test not batched input _snake_case : Tuple = image_processor(image_inputs[0], ["""semantic"""], return_tensors="""pt""" ).pixel_values _snake_case , _snake_case : Dict = self.image_processing_tester.get_expected_values(a_ ) self.assertEqual( encoded_images.shape, (1, self.image_processing_tester.num_channels, expected_height, expected_width), ) # Test batched _snake_case , _snake_case : Any = self.image_processing_tester.get_expected_values(a_, batched=a_ ) _snake_case : Tuple = image_processor( a_, ["""semantic"""] * len(a_ ), return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ), ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _snake_case : List[str] = prepare_image_inputs(self.image_processing_tester, equal_resolution=a_, numpify=a_ ) for image in image_inputs: self.assertIsInstance(a_, np.ndarray ) # Test not batched input _snake_case : Optional[int] = image_processor(image_inputs[0], ["""semantic"""], return_tensors="""pt""" ).pixel_values _snake_case , _snake_case : Any = self.image_processing_tester.get_expected_values(a_ ) self.assertEqual( encoded_images.shape, (1, self.image_processing_tester.num_channels, expected_height, expected_width), ) # Test batched _snake_case , _snake_case : Tuple = self.image_processing_tester.get_expected_values(a_, batched=a_ ) _snake_case : Optional[Any] = image_processor( a_, ["""semantic"""] * len(a_ ), return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ), ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case : Any = prepare_image_inputs(self.image_processing_tester, equal_resolution=a_, torchify=a_ ) for image in image_inputs: self.assertIsInstance(a_, torch.Tensor ) # Test not batched input _snake_case : Dict = image_processor(image_inputs[0], ["""semantic"""], return_tensors="""pt""" ).pixel_values _snake_case , _snake_case : Union[str, Any] = self.image_processing_tester.get_expected_values(a_ ) self.assertEqual( encoded_images.shape, (1, self.image_processing_tester.num_channels, expected_height, expected_width), ) # Test batched _snake_case , _snake_case : Union[str, Any] = self.image_processing_tester.get_expected_values(a_, batched=a_ ) _snake_case : str = image_processor( a_, ["""semantic"""] * len(a_ ), return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ), ) def UpperCamelCase_ ( self: Dict, a_: List[Any]=False, a_: int=False, a_: Union[str, Any]="np" ): '''simple docstring''' _snake_case : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # prepare image and target _snake_case : Dict = self.image_processing_tester.num_labels _snake_case : int = None _snake_case : Union[str, Any] = None _snake_case : Any = prepare_image_inputs(self.image_processing_tester, equal_resolution=a_ ) if with_segmentation_maps: _snake_case : str = num_labels if is_instance_map: _snake_case : Union[str, Any] = list(range(a_ ) ) * 2 _snake_case : List[Any] = dict(enumerate(a_ ) ) _snake_case : Any = [ np.random.randint(0, high * 2, (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": _snake_case : List[str] = [Image.fromarray(a_ ) for annotation in annotations] _snake_case : Dict = image_processor( a_, ["""semantic"""] * len(a_ ), a_, return_tensors="""pt""", instance_id_to_semantic_id=a_, pad_and_return_pixel_mask=a_, ) return inputs def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self: Any ): '''simple docstring''' def common(a_: List[Any]=False, a_: Any=None ): _snake_case : str = self.comm_get_image_processor_inputs( with_segmentation_maps=a_, is_instance_map=a_, segmentation_type=a_ ) _snake_case : str = inputs["""mask_labels"""] _snake_case : str = inputs["""class_labels"""] _snake_case : Optional[int] = inputs["""pixel_values"""] _snake_case : int = inputs["""text_inputs"""] # check the batch_size for mask_label, class_label, text_input in zip(a_, a_, a_ ): self.assertEqual(mask_label.shape[0], class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:], pixel_values.shape[2:] ) self.assertEqual(len(a_ ), self.image_processing_tester.num_text ) common() common(is_instance_map=a_ ) common(is_instance_map=a_, segmentation_type="""pil""" ) common(is_instance_map=a_, segmentation_type="""pil""" ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = np.zeros((20, 50) ) _snake_case : Any = 1 _snake_case : Dict = 1 _snake_case : Tuple = 1 _snake_case : List[Any] = binary_mask_to_rle(a_ ) self.assertEqual(len(a_ ), 4 ) self.assertEqual(rle[0], 21 ) self.assertEqual(rle[1], 45 ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = self.image_processing_class( num_labels=self.image_processing_tester.num_classes, max_seq_length=77, task_seq_length=77, class_info_file="""ade20k_panoptic.json""", num_text=self.image_processing_tester.num_text, repo_path="""shi-labs/oneformer_demo""", ) _snake_case : Dict = self.image_processing_tester.get_fake_oneformer_outputs() _snake_case : Union[str, Any] = fature_extractor.post_process_semantic_segmentation(a_ ) self.assertEqual(len(a_ ), self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape, ( self.image_processing_tester.height, self.image_processing_tester.width, ), ) _snake_case : List[Any] = [(1, 4) for i in range(self.image_processing_tester.batch_size )] _snake_case : Dict = fature_extractor.post_process_semantic_segmentation(a_, target_sizes=a_ ) self.assertEqual(segmentation[0].shape, target_sizes[0] ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = self.image_processing_class( num_labels=self.image_processing_tester.num_classes, max_seq_length=77, task_seq_length=77, class_info_file="""ade20k_panoptic.json""", num_text=self.image_processing_tester.num_text, repo_path="""shi-labs/oneformer_demo""", ) _snake_case : Union[str, Any] = self.image_processing_tester.get_fake_oneformer_outputs() _snake_case : str = image_processor.post_process_instance_segmentation(a_, threshold=0 ) self.assertTrue(len(a_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ), a_ ) self.assertEqual( el["""segmentation"""].shape, (self.image_processing_tester.height, self.image_processing_tester.width) ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : str = self.image_processing_class( num_labels=self.image_processing_tester.num_classes, max_seq_length=77, task_seq_length=77, class_info_file="""ade20k_panoptic.json""", num_text=self.image_processing_tester.num_text, repo_path="""shi-labs/oneformer_demo""", ) _snake_case : List[Any] = self.image_processing_tester.get_fake_oneformer_outputs() _snake_case : Optional[Any] = image_processor.post_process_panoptic_segmentation(a_, threshold=0 ) self.assertTrue(len(a_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ), a_ ) self.assertEqual( el["""segmentation"""].shape, (self.image_processing_tester.height, self.image_processing_tester.width) )
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" stooge(snake_case__ , 0 , len(snake_case__ ) - 1 ) return arr def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any , snake_case__ : int ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) # Recursively sort last 2/3 elements stooge(snake_case__ , i + t , (snake_case__) ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
64
1
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
64
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowercase( metaclass=__a ): '''simple docstring''' lowercase__ = ["note_seq"] def __init__( self: Dict, *a_: Union[str, Any], **a_: List[str] ): '''simple docstring''' requires_backends(self, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Optional[int], *a_: Any, **a_: Optional[Any] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Tuple, *a_: Optional[Any], **a_: List[str] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] )
64
1
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class lowercase( __a ): '''simple docstring''' lowercase__ = 42 lowercase__ = jnp.floataa lowercase__ = True def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setup() _snake_case : Optional[int] = nn.Dense(5, dtype=self.dtype ) def __call__( self: Optional[Any], *a_: Any, **a_: Any ): '''simple docstring''' _snake_case : Tuple = super().__call__(*a_, **a_ ) _snake_case : int = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class lowercase( __a ): '''simple docstring''' lowercase__ = FlaxBigBirdForNaturalQuestionsModule def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Tuple , snake_case__ : Tuple , snake_case__ : str , snake_case__ : Dict , snake_case__ : Any ): """simple docstring""" def cross_entropy(snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Dict=None ): _snake_case : List[str] = logits.shape[-1] _snake_case : Optional[Any] = (labels[..., None] == jnp.arange(snake_case__ )[None]).astype("""f4""" ) _snake_case : Optional[Any] = jax.nn.log_softmax(snake_case__ , axis=-1 ) _snake_case : Optional[int] = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: _snake_case : Tuple = reduction(snake_case__ ) return loss _snake_case : Union[str, Any] = partial(snake_case__ , reduction=jnp.mean ) _snake_case : List[Any] = cross_entropy(snake_case__ , snake_case__ ) _snake_case : Dict = cross_entropy(snake_case__ , snake_case__ ) _snake_case : int = cross_entropy(snake_case__ , snake_case__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class lowercase: '''simple docstring''' lowercase__ = "google/bigbird-roberta-base" lowercase__ = 30_00 lowercase__ = 1_05_00 lowercase__ = 1_28 lowercase__ = 3 lowercase__ = 1 lowercase__ = 5 # tx_args lowercase__ = 3e-5 lowercase__ = 0.0 lowercase__ = 2_00_00 lowercase__ = 0.0095 lowercase__ = "bigbird-roberta-natural-questions" lowercase__ = "training-expt" lowercase__ = "data/nq-training.jsonl" lowercase__ = "data/nq-validation.jsonl" def UpperCamelCase_ ( self: Dict ): '''simple docstring''' os.makedirs(self.base_dir, exist_ok=a_ ) _snake_case : Optional[Any] = os.path.join(self.base_dir, self.save_dir ) _snake_case : Dict = self.batch_size_per_device * jax.device_count() @dataclass class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = 40_96 # no dynamic padding on TPUs def __call__( self: List[Any], a_: Optional[int] ): '''simple docstring''' _snake_case : Optional[int] = self.collate_fn(a_ ) _snake_case : Optional[int] = jax.tree_util.tree_map(a_, a_ ) return batch def UpperCamelCase_ ( self: Optional[int], a_: Dict ): '''simple docstring''' _snake_case , _snake_case : Optional[int] = self.fetch_inputs(features["""input_ids"""] ) _snake_case : Union[str, Any] = { """input_ids""": jnp.array(a_, dtype=jnp.intaa ), """attention_mask""": jnp.array(a_, dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""], dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""], dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""], dtype=jnp.intaa ), } return batch def UpperCamelCase_ ( self: List[str], a_: list ): '''simple docstring''' _snake_case : List[str] = [self._fetch_inputs(a_ ) for ids in input_ids] return zip(*a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: list ): '''simple docstring''' _snake_case : Tuple = [1 for _ in range(len(a_ ) )] while len(a_ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Dict , snake_case__ : Tuple=None ): """simple docstring""" if seed is not None: _snake_case : List[Any] = dataset.shuffle(seed=snake_case__ ) for i in range(len(snake_case__ ) // batch_size ): _snake_case : Optional[int] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(snake_case__ ) @partial(jax.pmap , axis_name="""batch""" ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[Any] , **snake_case__ : str ): """simple docstring""" def loss_fn(snake_case__ : List[str] ): _snake_case : Dict = model_inputs.pop("""start_labels""" ) _snake_case : Optional[Any] = model_inputs.pop("""end_labels""" ) _snake_case : List[str] = model_inputs.pop("""pooled_labels""" ) _snake_case : Optional[int] = state.apply_fn(**snake_case__ , params=snake_case__ , dropout_rng=snake_case__ , train=snake_case__ ) _snake_case , _snake_case , _snake_case : Tuple = outputs return state.loss_fn( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) _snake_case , _snake_case : Union[str, Any] = jax.random.split(snake_case__ ) _snake_case : str = jax.value_and_grad(snake_case__ ) _snake_case , _snake_case : Tuple = grad_fn(state.params ) _snake_case : Dict = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) _snake_case : List[Any] = jax.lax.pmean(snake_case__ , """batch""" ) _snake_case : str = state.apply_gradients(grads=snake_case__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="""batch""" ) def UpperCAmelCase__ (snake_case__ : Any , **snake_case__ : Any ): """simple docstring""" _snake_case : Dict = model_inputs.pop("""start_labels""" ) _snake_case : List[Any] = model_inputs.pop("""end_labels""" ) _snake_case : List[str] = model_inputs.pop("""pooled_labels""" ) _snake_case : Tuple = state.apply_fn(**snake_case__ , params=state.params , train=snake_case__ ) _snake_case , _snake_case , _snake_case : Union[str, Any] = outputs _snake_case : Dict = state.loss_fn(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) _snake_case : List[Any] = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) return metrics class lowercase( train_state.TrainState ): '''simple docstring''' lowercase__ = struct.field(pytree_node=__a ) @dataclass class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = None def UpperCamelCase_ ( self: Dict, a_: Dict, a_: Tuple, a_: Optional[int], a_: Any=None ): '''simple docstring''' _snake_case : Optional[int] = model.params _snake_case : Tuple = TrainState.create( apply_fn=model.__call__, params=a_, tx=a_, loss_fn=a_, ) if ckpt_dir is not None: _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[Any] = restore_checkpoint(a_, a_ ) _snake_case : int = { """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } _snake_case , _snake_case : Optional[int] = build_tx(**a_ ) _snake_case : Union[str, Any] = train_state.TrainState( step=a_, apply_fn=model.__call__, params=a_, tx=a_, opt_state=a_, ) _snake_case : Optional[int] = args _snake_case : List[Any] = data_collator _snake_case : int = lr _snake_case : Optional[Any] = params _snake_case : Optional[int] = jax_utils.replicate(a_ ) return state def UpperCamelCase_ ( self: Union[str, Any], a_: Union[str, Any], a_: str, a_: Dict ): '''simple docstring''' _snake_case : List[Any] = self.args _snake_case : Optional[int] = len(a_ ) // args.batch_size _snake_case : Tuple = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(a_, jax.device_count() ) for epoch in range(args.max_epochs ): _snake_case : Optional[int] = jnp.array(0, dtype=jnp.floataa ) _snake_case : Dict = get_batched_dataset(a_, args.batch_size, seed=a_ ) _snake_case : Optional[int] = 0 for batch in tqdm(a_, total=a_, desc=f"Running EPOCH-{epoch}" ): _snake_case : str = self.data_collator(a_ ) _snake_case , _snake_case , _snake_case : Optional[Any] = self.train_step_fn(a_, a_, **a_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: _snake_case : Optional[Any] = jax_utils.unreplicate(state.step ) _snake_case : List[str] = running_loss.item() / i _snake_case : Tuple = self.scheduler_fn(state_step - 1 ) _snake_case : Tuple = self.evaluate(a_, a_ ) _snake_case : Union[str, Any] = { """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(a_ ) ) self.logger.log(a_, commit=a_ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f"-e{epoch}-s{i}", state=a_ ) def UpperCamelCase_ ( self: Tuple, a_: Optional[int], a_: Optional[Any] ): '''simple docstring''' _snake_case : str = get_batched_dataset(a_, self.args.batch_size ) _snake_case : Any = len(a_ ) // self.args.batch_size _snake_case : Union[str, Any] = jnp.array(0, dtype=jnp.floataa ) _snake_case : str = 0 for batch in tqdm(a_, total=a_, desc="""Evaluating ... """ ): _snake_case : List[Any] = self.data_collator(a_ ) _snake_case : str = self.val_step_fn(a_, **a_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def UpperCamelCase_ ( self: Union[str, Any], a_: List[str], a_: Optional[Any] ): '''simple docstring''' _snake_case : Optional[int] = jax_utils.unreplicate(a_ ) print(f"SAVING CHECKPOINT IN {save_dir}", end=""" ... """ ) self.model_save_fn(a_, params=state.params ) with open(os.path.join(a_, """opt_state.msgpack""" ), """wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args, os.path.join(a_, """args.joblib""" ) ) joblib.dump(self.data_collator, os.path.join(a_, """data_collator.joblib""" ) ) with open(os.path.join(a_, """training_state.json""" ), """w""" ) as f: json.dump({"""step""": state.step.item()}, a_ ) print("""DONE""" ) def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Dict ): """simple docstring""" print(F"RESTORING CHECKPOINT FROM {save_dir}" , end=""" ... """ ) with open(os.path.join(snake_case__ , """flax_model.msgpack""" ) , """rb""" ) as f: _snake_case : Tuple = from_bytes(state.params , f.read() ) with open(os.path.join(snake_case__ , """opt_state.msgpack""" ) , """rb""" ) as f: _snake_case : Any = from_bytes(state.opt_state , f.read() ) _snake_case : str = joblib.load(os.path.join(snake_case__ , """args.joblib""" ) ) _snake_case : Union[str, Any] = joblib.load(os.path.join(snake_case__ , """data_collator.joblib""" ) ) with open(os.path.join(snake_case__ , """training_state.json""" ) , """r""" ) as f: _snake_case : Tuple = json.load(snake_case__ ) _snake_case : int = training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : Any ): """simple docstring""" _snake_case : Dict = num_train_steps - warmup_steps _snake_case : Tuple = optax.linear_schedule(init_value=snake_case__ , end_value=snake_case__ , transition_steps=snake_case__ ) _snake_case : Tuple = optax.linear_schedule(init_value=snake_case__ , end_value=1e-7 , transition_steps=snake_case__ ) _snake_case : List[str] = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : Optional[Any] , snake_case__ : Union[str, Any] , snake_case__ : Tuple ): """simple docstring""" def weight_decay_mask(snake_case__ : Tuple ): _snake_case : List[str] = traverse_util.flatten_dict(snake_case__ ) _snake_case : List[Any] = {k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(snake_case__ ) _snake_case : List[str] = scheduler_fn(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) _snake_case : Any = optax.adamw(learning_rate=snake_case__ , weight_decay=snake_case__ , mask=snake_case__ ) return tx, lr
64
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: List[str] ): '''simple docstring''' _snake_case : int = data _snake_case : Dict = [0X67452301, 0Xefcdab89, 0X98badcfe, 0X10325476, 0Xc3d2e1f0] @staticmethod def UpperCamelCase_ ( a_: Optional[Any], a_: Dict ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0Xffffffff def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _snake_case : Optional[int] = self.data + padding + struct.pack(""">Q""", 8 * len(self.data ) ) return padded_data def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0, len(self.padded_data ), 64 ) ] def UpperCamelCase_ ( self: Optional[Any], a_: List[Any] ): '''simple docstring''' _snake_case : List[str] = list(struct.unpack(""">16L""", a_ ) ) + [0] * 64 for i in range(16, 80 ): _snake_case : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]), 1 ) return w def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = self.padding() _snake_case : str = self.split_blocks() for block in self.blocks: _snake_case : Any = self.expand_block(a_ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = self.h for i in range(0, 80 ): if 0 <= i < 20: _snake_case : int = (b & c) | ((~b) & d) _snake_case : str = 0X5a827999 elif 20 <= i < 40: _snake_case : Optional[int] = b ^ c ^ d _snake_case : str = 0X6ed9eba1 elif 40 <= i < 60: _snake_case : List[Any] = (b & c) | (b & d) | (c & d) _snake_case : List[Any] = 0X8f1bbcdc elif 60 <= i < 80: _snake_case : List[Any] = b ^ c ^ d _snake_case : int = 0Xca62c1d6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = ( self.rotate(a_, 5 ) + f + e + k + expanded_block[i] & 0Xffffffff, a, self.rotate(a_, 30 ), c, d, ) _snake_case : Union[str, Any] = ( self.h[0] + a & 0Xffffffff, self.h[1] + b & 0Xffffffff, self.h[2] + c & 0Xffffffff, self.h[3] + d & 0Xffffffff, self.h[4] + e & 0Xffffffff, ) return ("{:08x}" * 5).format(*self.h ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = B"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) _snake_case : Union[str, Any] = parser.parse_args() _snake_case : List[Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: _snake_case : str = f.read() else: _snake_case : int = bytes(snake_case__ , """utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
64
1
"""simple docstring""" import contextlib from multiprocessing import Pool, RLock from tqdm.auto import tqdm from ..utils import experimental, logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = None @experimental def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Any , snake_case__ : Optional[int] , snake_case__ : Tuple , snake_case__ : str , snake_case__ : Optional[Any] , snake_case__ : int ): """simple docstring""" if ParallelBackendConfig.backend_name is None: return _map_with_multiprocessing_pool( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) return _map_with_joblib(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : Any , snake_case__ : Tuple , snake_case__ : Optional[Any] , snake_case__ : Optional[int] , snake_case__ : int , snake_case__ : Tuple ): """simple docstring""" _snake_case : str = num_proc if num_proc <= len(snake_case__ ) else len(snake_case__ ) _snake_case : Dict = [] # We organize the splits ourselve (contiguous splits) for index in range(snake_case__ ): _snake_case : str = len(snake_case__ ) // num_proc _snake_case : Tuple = len(snake_case__ ) % num_proc _snake_case : Tuple = div * index + min(snake_case__ , snake_case__ ) _snake_case : str = start + div + (1 if index < mod else 0) split_kwds.append((function, iterable[start:end], types, index, disable_tqdm, desc) ) if len(snake_case__ ) != sum(len(i[1] ) for i in split_kwds ): raise ValueError( F"Error dividing inputs iterable among processes. " F"Total number of objects {len(snake_case__ )}, " F"length: {sum(len(i[1] ) for i in split_kwds )}" ) logger.info( F"Spawning {num_proc} processes for {len(snake_case__ )} objects in slices of {[len(i[1] ) for i in split_kwds]}" ) _snake_case , _snake_case : List[str] = None, None if not disable_tqdm: _snake_case , _snake_case : List[Any] = (RLock(),), tqdm.set_lock with Pool(snake_case__ , initargs=snake_case__ , initializer=snake_case__ ) as pool: _snake_case : List[Any] = pool.map(snake_case__ , snake_case__ ) logger.info(F"Finished {num_proc} processes" ) _snake_case : List[Any] = [obj for proc_res in mapped for obj in proc_res] logger.info(F"Unpacked {len(snake_case__ )} objects" ) return mapped def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : Union[str, Any] , snake_case__ : List[Any] ): """simple docstring""" import joblib with joblib.parallel_backend(ParallelBackendConfig.backend_name , n_jobs=snake_case__ ): return joblib.Parallel()( joblib.delayed(snake_case__ )((function, obj, types, None, True, None) ) for obj in iterable ) @experimental @contextlib.contextmanager def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = backend_name if backend_name == "spark": from joblibspark import register_spark register_spark() # TODO: call create_cache_and_write_probe if "download" in steps # TODO: raise NotImplementedError when Dataset.map etc is called try: yield finally: _snake_case : Tuple = None
64
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
1
"""simple docstring""" from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class lowercase: '''simple docstring''' def UpperCamelCase_ ( self: Tuple, a_: Optional[int] ): '''simple docstring''' raise NotImplementedError() def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' raise NotImplementedError() class lowercase( __a ): '''simple docstring''' def __init__( self: Optional[Any], a_: "AutoTokenizer", a_: bool = False, **a_: Optional[int] ): '''simple docstring''' _snake_case : Dict = tokenizer _snake_case : Optional[int] = skip_prompt _snake_case : Optional[Any] = decode_kwargs # variables used in the streaming process _snake_case : Optional[int] = [] _snake_case : Optional[Any] = 0 _snake_case : Union[str, Any] = True def UpperCamelCase_ ( self: Optional[int], a_: Union[str, Any] ): '''simple docstring''' if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("""TextStreamer only supports batch size 1""" ) elif len(value.shape ) > 1: _snake_case : List[Any] = value[0] if self.skip_prompt and self.next_tokens_are_prompt: _snake_case : Optional[int] = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) _snake_case : int = self.tokenizer.decode(self.token_cache, **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("""\n""" ): _snake_case : Optional[int] = text[self.print_len :] _snake_case : Optional[Any] = [] _snake_case : Optional[Any] = 0 # If the last token is a CJK character, we print the characters. elif len(a_ ) > 0 and self._is_chinese_char(ord(text[-1] ) ): _snake_case : Optional[int] = text[self.print_len :] self.print_len += len(a_ ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: _snake_case : List[Any] = text[self.print_len : text.rfind(""" """ ) + 1] self.print_len += len(a_ ) self.on_finalized_text(a_ ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' if len(self.token_cache ) > 0: _snake_case : Dict = self.tokenizer.decode(self.token_cache, **self.decode_kwargs ) _snake_case : Tuple = text[self.print_len :] _snake_case : List[Any] = [] _snake_case : Dict = 0 else: _snake_case : str = """""" _snake_case : Tuple = True self.on_finalized_text(a_, stream_end=a_ ) def UpperCamelCase_ ( self: Any, a_: str, a_: bool = False ): '''simple docstring''' print(a_, flush=a_, end="""""" if not stream_end else None ) def UpperCamelCase_ ( self: Any, a_: int ): '''simple docstring''' if ( (cp >= 0X4e00 and cp <= 0X9fff) or (cp >= 0X3400 and cp <= 0X4dbf) # or (cp >= 0X20000 and cp <= 0X2a6df) # or (cp >= 0X2a700 and cp <= 0X2b73f) # or (cp >= 0X2b740 and cp <= 0X2b81f) # or (cp >= 0X2b820 and cp <= 0X2ceaf) # or (cp >= 0Xf900 and cp <= 0Xfaff) or (cp >= 0X2f800 and cp <= 0X2fa1f) # ): # return True return False class lowercase( __a ): '''simple docstring''' def __init__( self: str, a_: "AutoTokenizer", a_: bool = False, a_: Optional[float] = None, **a_: List[str] ): '''simple docstring''' super().__init__(a_, a_, **a_ ) _snake_case : Any = Queue() _snake_case : Optional[int] = None _snake_case : List[Any] = timeout def UpperCamelCase_ ( self: int, a_: str, a_: bool = False ): '''simple docstring''' self.text_queue.put(a_, timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal, timeout=self.timeout ) def __iter__( self: List[Any] ): '''simple docstring''' return self def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Any = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
64
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : list[int] ): """simple docstring""" if not len(snake_case__ ) == len(snake_case__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _snake_case , _snake_case , _snake_case : Optional[Any] = equationa _snake_case , _snake_case , _snake_case : Optional[Any] = equationa # Calculate the determinants of the matrices _snake_case : Any = aa * ba - aa * ba _snake_case : Optional[int] = ca * ba - ca * ba _snake_case : Tuple = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _snake_case : int = determinant_x / determinant _snake_case : List[str] = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list ): """simple docstring""" if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] _snake_case : List[Any] = [] def generate(snake_case__ : int , snake_case__ : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _snake_case , _snake_case : Optional[Any] = arr[k - 1], arr[i] else: # k is odd _snake_case , _snake_case : List[str] = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
64
1
"""simple docstring""" # using dfs for finding eulerian path traversal def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : Tuple , snake_case__ : int , snake_case__ : List[str]=None ): """simple docstring""" _snake_case : List[Any] = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: _snake_case , _snake_case : Dict = True, True _snake_case : str = dfs(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) return path def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : List[str] = 0 _snake_case : List[str] = -1 for i in range(snake_case__ ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 _snake_case : int = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] _snake_case , _snake_case : Dict = check_circuit_or_path(snake_case__ , snake_case__ ) if check == 3: print("""graph is not Eulerian""" ) print("""no path""" ) return _snake_case : int = 1 if check == 2: _snake_case : Optional[int] = odd_node print("""graph has a Euler path""" ) if check == 1: print("""graph has a Euler cycle""" ) _snake_case : Optional[int] = dfs(snake_case__ , snake_case__ , snake_case__ ) print(snake_case__ ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} _snake_case : Dict = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} _snake_case : Optional[Any] = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} _snake_case : List[str] = {1: [2, 3], 2: [1, 3], 3: [1, 2]} _snake_case : List[str] = { 1: [], 2: [] # all degree is zero } _snake_case : List[Any] = 10 check_euler(snake_case__ , snake_case__ ) check_euler(snake_case__ , snake_case__ ) check_euler(snake_case__ , snake_case__ ) check_euler(snake_case__ , snake_case__ ) check_euler(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
"""simple docstring""" from math import factorial A_ = {str(d): factorial(d) for d in range(10)} def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" return sum(DIGIT_FACTORIAL[d] for d in str(snake_case__ ) ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , snake_case__ ) if sum_of_digit_factorial(snake_case__ ) == i ) if __name__ == "__main__": print(F'''{solution() = }''')
64
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" if len(snake_case__ ) < k or k < 0: raise ValueError("""Invalid Input""" ) _snake_case : Optional[int] = sum(array[:k] ) for i in range(len(snake_case__ ) - k ): _snake_case : Optional[Any] = current_sum - array[i] + array[i + k] _snake_case : List[str] = max(snake_case__ , snake_case__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-10_00, 10_00) for i in range(1_00)] A_ = randint(0, 1_10) print(F'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
64
1
"""simple docstring""" import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration A_ = 50_00_00 A_ , A_ = os.path.split(__file__) A_ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def UpperCAmelCase__ (snake_case__ : datasets.Dataset , **snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Tuple = dataset.map(**snake_case__ ) @get_duration def UpperCAmelCase__ (snake_case__ : datasets.Dataset , **snake_case__ : Any ): """simple docstring""" _snake_case : List[str] = dataset.filter(**snake_case__ ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Dict = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : Dict = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) _snake_case : List[Any] = generate_example_dataset( os.path.join(snake_case__ , """dataset.arrow""" ) , snake_case__ , num_examples=snake_case__ ) _snake_case : List[Any] = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=snake_case__ ) def tokenize(snake_case__ : Optional[int] ): return tokenizer(examples["""text"""] ) _snake_case : str = map(snake_case__ ) _snake_case : Optional[int] = map(snake_case__ , batched=snake_case__ ) _snake_case : int = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""numpy""" ): _snake_case : Dict = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""pandas""" ): _snake_case : List[str] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""torch""" , columns="""numbers""" ): _snake_case : Union[str, Any] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""tensorflow""" , columns="""numbers""" ): _snake_case : List[str] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) _snake_case : Dict = map(snake_case__ , function=snake_case__ , batched=snake_case__ ) _snake_case : List[str] = filter(snake_case__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(snake_case__ , """wb""" ) as f: f.write(json.dumps(snake_case__ ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class lowercase( __a ): '''simple docstring''' def __init__( self: Optional[int], a_: int=None, **a_: Union[str, Any] ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""", a_, ) super().__init__(args=a_, **a_ )
64
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = CodeGenTokenizer lowercase__ = CodeGenTokenizerFast lowercase__ = True lowercase__ = {"add_prefix_space": True} lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : Tuple = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] _snake_case : Tuple = dict(zip(a_, range(len(a_ ) ) ) ) _snake_case : str = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _snake_case : List[Any] = {"""unk_token""": """<unk>"""} _snake_case : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(a_ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(a_ ) ) def UpperCamelCase_ ( self: Any, **a_: int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Any, **a_: str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Dict ): '''simple docstring''' _snake_case : Union[str, Any] = """lower newer""" _snake_case : Tuple = """lower newer""" return input_text, output_text def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = CodeGenTokenizer(self.vocab_file, self.merges_file, **self.special_tokens_map ) _snake_case : Optional[Any] = """lower newer""" _snake_case : Optional[int] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _snake_case : int = tokenizer.tokenize(a_, add_prefix_space=a_ ) self.assertListEqual(a_, a_ ) _snake_case : str = tokens + [tokenizer.unk_token] _snake_case : Optional[int] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : int = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : Dict = """lower newer""" # Testing tokenization _snake_case : Dict = tokenizer.tokenize(a_, add_prefix_space=a_ ) _snake_case : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids without special tokens _snake_case : Optional[Any] = tokenizer.encode(a_, add_special_tokens=a_, add_prefix_space=a_ ) _snake_case : Tuple = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids with special tokens _snake_case : Tuple = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : int = tokenizer.encode(a_, add_prefix_space=a_ ) _snake_case : Optional[Any] = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) # Testing the unknown token _snake_case : Tuple = tokens + [rust_tokenizer.unk_token] _snake_case : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: List[Any]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) # Simple input _snake_case : Any = """This is a simple input""" _snake_case : Optional[int] = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") _snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) # Pair input self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname, pad_token="""<pad>""" ) # Simple input _snake_case : List[Any] = """This is a simple input""" _snake_case : int = ["""This is a simple input looooooooong""", """This is a simple input"""] _snake_case : Any = ("""This is a simple input""", """This is a pair""") _snake_case : str = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] _snake_case : str = tokenizer.pad_token_id _snake_case : Optional[int] = tokenizer(a_, padding="""max_length""", max_length=30, return_tensors="""np""" ) _snake_case : Dict = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) _snake_case : Tuple = tokenizer(*a_, padding="""max_length""", max_length=60, return_tensors="""np""" ) _snake_case : Optional[Any] = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1], 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1], 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1], 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1], 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = """$$$""" _snake_case : List[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname, bos_token=a_, add_bos_token=a_ ) _snake_case : str = """This is a simple input""" _snake_case : int = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Tuple = tokenizer(a_ ) _snake_case : Optional[Any] = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0], a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _snake_case : Optional[int] = tokenizer.decode(out_s.input_ids ) _snake_case : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0], a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) _snake_case : Dict = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" _snake_case : Union[str, Any] = """\nif len_a > len_b: result = a\nelse: result = b""" _snake_case : Optional[Any] = tokenizer.encode(a_ ) _snake_case : Dict = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] _snake_case : Optional[Any] = tokenizer.decode(a_, truncate_before_pattern=a_ ) self.assertEqual(a_, a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : str , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = """""" for word_or_phrase in separated: if not isinstance(snake_case__ , snake_case__ ): raise Exception("""join() accepts only strings to be joined""" ) joined += word_or_phrase + separator return joined.strip(snake_case__ ) if __name__ == "__main__": from doctest import testmod testmod()
64
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ = re.compile(r'''\s+''') def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" return {"hash": hashlib.mda(re.sub(snake_case__ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : Any = [len(snake_case__ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(snake_case__ ), "line_max": max(snake_case__ )} def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : List[Any] ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : List[str]=5 ): """simple docstring""" _snake_case : Any = ["""auto-generated""", """autogenerated""", """automatically generated"""] _snake_case : Tuple = example["""content"""].splitlines() for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Union[str, Any]=5 , snake_case__ : Any=0.05 ): """simple docstring""" _snake_case : Optional[Any] = ["""unit tests""", """test file""", """configuration file"""] _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : Dict = 0 _snake_case : str = 0 # first test for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _snake_case : Optional[int] = example["""content"""].count("""\n""" ) _snake_case : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[int] = ["""def """, """class """, """for """, """while """] _snake_case : str = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : List[str]=4 ): """simple docstring""" _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : str = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" _snake_case : Optional[Any] = tokenizer(example["""content"""] , truncation=snake_case__ )["""input_ids"""] _snake_case : Optional[Any] = len(example["""content"""] ) / len(snake_case__ ) return {"ratio": ratio} def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = {} results.update(get_hash(snake_case__ ) ) results.update(line_stats(snake_case__ ) ) results.update(alpha_stats(snake_case__ ) ) results.update(char_token_ratio(snake_case__ ) ) results.update(is_autogenerated(snake_case__ ) ) results.update(is_config_or_test(snake_case__ ) ) results.update(has_no_keywords(snake_case__ ) ) results.update(has_few_assignments(snake_case__ ) ) return results def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] ): """simple docstring""" if not check_uniques(snake_case__ , snake_case__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """rb""" ) as f_in: with gzip.open(str(snake_case__ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(snake_case__ , snake_case__ ) os.unlink(snake_case__ ) # Settings A_ = HfArgumentParser(PreprocessingArguments) A_ = parser.parse_args() if args.num_workers is None: A_ = multiprocessing.cpu_count() A_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ = time.time() A_ = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ = time.time() A_ = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ = set(ds.unique('''hash''')) A_ = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ = time.time() A_ = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ = time.time() A_ , A_ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) A_ = output_dir / '''data''' data_dir.mkdir(exist_ok=True) A_ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ = str(data_dir / F'''file-{file_number+1:012}.json''') A_ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
64
1
"""simple docstring""" import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version A_ = logging.getLogger(__name__) require_version('''pytorch_lightning>=1.0.4''') A_ = { '''base''': AutoModel, '''sequence-classification''': AutoModelForSequenceClassification, '''question-answering''': AutoModelForQuestionAnswering, '''pretraining''': AutoModelForPreTraining, '''token-classification''': AutoModelForTokenClassification, '''language-modeling''': AutoModelWithLMHead, '''summarization''': AutoModelForSeqaSeqLM, '''translation''': AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization A_ = { '''linear''': get_linear_schedule_with_warmup, '''cosine''': get_cosine_schedule_with_warmup, '''cosine_w_restarts''': get_cosine_with_hard_restarts_schedule_with_warmup, '''polynomial''': get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } A_ = sorted(arg_to_scheduler.keys()) A_ = '''{''' + ''', '''.join(arg_to_scheduler_choices) + '''}''' class lowercase( pl.LightningModule ): '''simple docstring''' def __init__( self: List[str], a_: argparse.Namespace, a_: Tuple=None, a_: Dict="base", a_: Tuple=None, a_: Union[str, Any]=None, a_: int=None, **a_: int, ): '''simple docstring''' super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(a_ ) _snake_case : Any = 0 _snake_case : str = Path(self.hparams.output_dir ) _snake_case : List[Any] = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: _snake_case : Optional[int] = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path, **({"""num_labels""": num_labels} if num_labels is not None else {}), cache_dir=a_, **a_, ) else: _snake_case : PretrainedConfig = config _snake_case : Dict = ("""encoder_layerdrop""", """decoder_layerdrop""", """dropout""", """attention_dropout""") for p in extra_model_params: if getattr(self.hparams, a_, a_ ): assert hasattr(self.config, a_ ), f"model config doesn't have a `{p}` attribute" setattr(self.config, a_, getattr(self.hparams, a_ ) ) if tokenizer is None: _snake_case : Union[str, Any] = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path, cache_dir=a_, ) else: _snake_case : PreTrainedTokenizer = tokenizer _snake_case : Union[str, Any] = MODEL_MODES[mode] if model is None: _snake_case : Dict = self.model_type.from_pretrained( self.hparams.model_name_or_path, from_tf=bool(""".ckpt""" in self.hparams.model_name_or_path ), config=self.config, cache_dir=a_, ) else: _snake_case : int = model def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: List[str] ): '''simple docstring''' _snake_case : List[Any] = self.model_type.from_pretrained(*a_, **a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Dict = arg_to_scheduler[self.hparams.lr_scheduler] _snake_case : Optional[int] = get_schedule_func( self.opt, num_warmup_steps=self.hparams.warmup_steps, num_training_steps=self.total_steps() ) _snake_case : Dict = {"""scheduler""": scheduler, """interval""": """step""", """frequency""": 1} return scheduler def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Any = self.model _snake_case : List[Any] = ["""bias""", """LayerNorm.weight"""] _snake_case : Any = [ { """params""": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters """weight_decay""": self.hparams.weight_decay, }, { """params""": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] if self.hparams.adafactor: _snake_case : Optional[int] = Adafactor( a_, lr=self.hparams.learning_rate, scale_parameter=a_, relative_step=a_ ) else: _snake_case : Optional[int] = AdamW( a_, lr=self.hparams.learning_rate, eps=self.hparams.adam_epsilon ) _snake_case : Optional[Any] = optimizer _snake_case : Tuple = self.get_lr_scheduler() return [optimizer], [scheduler] def UpperCamelCase_ ( self: int, a_: int, a_: Optional[Any] ): '''simple docstring''' return self.validation_step(a_, a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Union[str, Any] ): '''simple docstring''' return self.validation_end(a_ ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : List[Any] = max(1, self.hparams.gpus ) # TODO: consider num_tpu_cores _snake_case : Optional[Any] = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def UpperCamelCase_ ( self: Optional[Any], a_: Optional[Any] ): '''simple docstring''' if stage == "test": _snake_case : Union[str, Any] = len(self.test_dataloader().dataset ) else: _snake_case : Union[str, Any] = self.get_dataloader("""train""", self.hparams.train_batch_size, shuffle=a_ ) _snake_case : Union[str, Any] = len(self.train_dataloader().dataset ) def UpperCamelCase_ ( self: List[Any], a_: str, a_: int, a_: bool = False ): '''simple docstring''' raise NotImplementedError("""You must implement this for your task""" ) def UpperCamelCase_ ( self: str ): '''simple docstring''' return self.train_loader def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return self.get_dataloader("""dev""", self.hparams.eval_batch_size, shuffle=a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return self.get_dataloader("""test""", self.hparams.eval_batch_size, shuffle=a_ ) def UpperCamelCase_ ( self: Dict, a_: int ): '''simple docstring''' return os.path.join( self.hparams.data_dir, """cached_{}_{}_{}""".format( a_, list(filter(a_, self.hparams.model_name_or_path.split("""/""" ) ) ).pop(), str(self.hparams.max_seq_length ), ), ) @pl.utilities.rank_zero_only def UpperCamelCase_ ( self: str, a_: Dict[str, Any] ): '''simple docstring''' _snake_case : Optional[int] = self.output_dir.joinpath("""best_tfmr""" ) _snake_case : Optional[int] = self.step_count self.model.save_pretrained(a_ ) self.tokenizer.save_pretrained(a_ ) @staticmethod def UpperCamelCase_ ( a_: Dict, a_: Dict ): '''simple docstring''' parser.add_argument( """--model_name_or_path""", default=a_, type=a_, required=a_, help="""Path to pretrained model or model identifier from huggingface.co/models""", ) parser.add_argument( """--config_name""", default="""""", type=a_, help="""Pretrained config name or path if not the same as model_name""" ) parser.add_argument( """--tokenizer_name""", default=a_, type=a_, help="""Pretrained tokenizer name or path if not the same as model_name""", ) parser.add_argument( """--cache_dir""", default=str(Path(a_ ).parent / """test_run""" / """cache""" ), type=a_, help="""Where do you want to store the pre-trained models downloaded from huggingface.co""", ) parser.add_argument( """--encoder_layerdrop""", type=a_, help="""Encoder layer dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--decoder_layerdrop""", type=a_, help="""Decoder layer dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--dropout""", type=a_, help="""Dropout probability (Optional). Goes into model.config""", ) parser.add_argument( """--attention_dropout""", type=a_, help="""Attention dropout probability (Optional). Goes into model.config""", ) parser.add_argument("""--learning_rate""", default=5E-5, type=a_, help="""The initial learning rate for Adam.""" ) parser.add_argument( """--lr_scheduler""", default="""linear""", choices=a_, metavar=a_, type=a_, help="""Learning rate scheduler""", ) parser.add_argument("""--weight_decay""", default=0.0, type=a_, help="""Weight decay if we apply some.""" ) parser.add_argument("""--adam_epsilon""", default=1E-8, type=a_, help="""Epsilon for Adam optimizer.""" ) parser.add_argument("""--warmup_steps""", default=0, type=a_, help="""Linear warmup over warmup_steps.""" ) parser.add_argument("""--num_workers""", default=4, type=a_, help="""kwarg passed to DataLoader""" ) parser.add_argument("""--num_train_epochs""", dest="""max_epochs""", default=3, type=a_ ) parser.add_argument("""--train_batch_size""", default=32, type=a_ ) parser.add_argument("""--eval_batch_size""", default=32, type=a_ ) parser.add_argument("""--adafactor""", action="""store_true""" ) class lowercase( pl.Callback ): '''simple docstring''' def UpperCamelCase_ ( self: Optional[int], a_: Optional[int], a_: Any ): '''simple docstring''' if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class lowercase( pl.Callback ): '''simple docstring''' def UpperCamelCase_ ( self: str, a_: Optional[Any], a_: List[str] ): '''simple docstring''' for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(a_ ) class lowercase( pl.Callback ): '''simple docstring''' def UpperCamelCase_ ( self: List[Any], a_: Union[str, Any], a_: str ): '''simple docstring''' _snake_case : Any = trainer.lr_schedulers[0]["""scheduler"""] _snake_case : List[str] = {f"lr_group_{i}": lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(a_ ) def UpperCamelCase_ ( self: Dict, a_: pl.Trainer, a_: pl.LightningModule ): '''simple docstring''' rank_zero_info("""***** Validation results *****""" ) _snake_case : List[str] = trainer.callback_metrics # Log results for key in sorted(a_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(a_, str(metrics[key] ) ) ) def UpperCamelCase_ ( self: Any, a_: pl.Trainer, a_: pl.LightningModule ): '''simple docstring''' rank_zero_info("""***** Test results *****""" ) _snake_case : Union[str, Any] = trainer.callback_metrics # Log and save results to file _snake_case : str = os.path.join(pl_module.hparams.output_dir, """test_results.txt""" ) with open(a_, """w""" ) as writer: for key in sorted(a_ ): if key not in ["log", "progress_bar"]: rank_zero_info("""{} = {}\n""".format(a_, str(metrics[key] ) ) ) writer.write("""{} = {}\n""".format(a_, str(metrics[key] ) ) ) def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : Optional[int] ): """simple docstring""" parser.add_argument( """--output_dir""" , default=str(Path(snake_case__ ).parent / """test_run""" / """model_checkpoints""" ) , type=snake_case__ , help="""The output directory where the model predictions and checkpoints will be written.""" , ) parser.add_argument( """--fp16""" , action="""store_true""" , help="""Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit""" , ) parser.add_argument( """--fp16_opt_level""" , type=snake_case__ , default="""O2""" , help=( """For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3'].""" """See details at https://nvidia.github.io/apex/amp.html""" ) , ) parser.add_argument("""--n_tpu_cores""" , dest="""tpu_cores""" , type=snake_case__ ) parser.add_argument("""--max_grad_norm""" , dest="""gradient_clip_val""" , default=1.0 , type=snake_case__ , help="""Max gradient norm""" ) parser.add_argument("""--do_train""" , action="""store_true""" , help="""Whether to run training.""" ) parser.add_argument("""--do_predict""" , action="""store_true""" , help="""Whether to run predictions on the test set.""" ) parser.add_argument( """--gradient_accumulation_steps""" , dest="""accumulate_grad_batches""" , type=snake_case__ , default=1 , help="""Number of updates steps to accumulate before performing a backward/update pass.""" , ) parser.add_argument("""--seed""" , type=snake_case__ , default=42 , help="""random seed for initialization""" ) parser.add_argument( """--data_dir""" , default=str(Path(snake_case__ ).parent / """test_run""" / """dummy-train-data""" ) , type=snake_case__ , help="""The input data dir. Should contain the training files for the CoNLL-2003 NER task.""" , ) def UpperCAmelCase__ (snake_case__ : BaseTransformer , snake_case__ : argparse.Namespace , snake_case__ : int=None , snake_case__ : str=True , snake_case__ : Optional[Any]=[] , snake_case__ : List[str]=None , snake_case__ : Any=None , **snake_case__ : Tuple , ): """simple docstring""" pl.seed_everything(args.seed ) # init model _snake_case : str = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=snake_case__ ) # add custom checkpoints if checkpoint_callback is None: _snake_case : Optional[Any] = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="""checkpoint""" , monitor="""val_loss""" , mode="""min""" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(snake_case__ ) if logging_callback is None: _snake_case : Union[str, Any] = LoggingCallback() _snake_case : List[str] = {} if args.fpaa: _snake_case : Union[str, Any] = 16 if args.gpus > 1: _snake_case : Optional[Any] = """auto""" _snake_case : Union[str, Any] = """ddp""" _snake_case : List[str] = args.accumulate_grad_batches _snake_case : Dict = None _snake_case : str = """auto""" _snake_case : Dict = pl.Trainer.from_argparse_args( snake_case__ , weights_summary=snake_case__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=snake_case__ , val_check_interval=1 , num_sanity_val_steps=2 , **snake_case__ , ) if args.do_train: trainer.fit(snake_case__ ) else: print("""RAG modeling tests with new set functions successfuly executed!""" ) return trainer
64
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
1
"""simple docstring""" from string import ascii_lowercase, ascii_uppercase def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if not sentence: return "" _snake_case : str = dict(zip(snake_case__ , snake_case__ ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
64
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : str , snake_case__ : str ): """simple docstring""" for attribute in key.split(""".""" ): _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ).shape else: _snake_case : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": _snake_case : int = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : Tuple = value elif weight_type == "bias": _snake_case : List[str] = value else: _snake_case : int = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str] ): """simple docstring""" _snake_case : List[Any] = [] _snake_case : Optional[Any] = fairseq_model.state_dict() _snake_case : str = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _snake_case : Optional[Any] = None for name, value in fairseq_dict.items(): _snake_case : Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) _snake_case : Dict = True elif name.split(""".""" )[0] == "proj": _snake_case : Dict = fairseq_model.proj _snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: _snake_case : Dict = True if "*" in mapped_key: _snake_case : Optional[int] = name.split(snake_case__ )[0].split(""".""" )[-2] _snake_case : Union[str, Any] = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: _snake_case : str = """weight_g""" elif "weight_v" in name: _snake_case : Optional[Any] = """weight_v""" elif "bias" in name: _snake_case : Union[str, Any] = """bias""" elif "weight" in name: _snake_case : int = """weight""" else: _snake_case : Optional[int] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Dict , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] , snake_case__ : int ): """simple docstring""" _snake_case : Any = full_name.split("""conv_layers.""" )[-1] _snake_case : Optional[int] = name.split(""".""" ) _snake_case : List[str] = int(items[0] ) _snake_case : Dict = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _snake_case : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _snake_case : int = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _snake_case : List[str] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" _snake_case , _snake_case : Optional[Any] = emb.weight.shape _snake_case : Optional[int] = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) _snake_case : Union[str, Any] = emb.weight.data return lin_layer def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Any = f.readlines() _snake_case : Optional[Any] = [line.split(""" """ )[0] for line in lines] _snake_case : str = len(snake_case__ ) _snake_case : Tuple = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Union[str, Any] , ): """simple docstring""" _snake_case : Optional[int] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : List[str] = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) _snake_case : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) _snake_case , _snake_case , _snake_case : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) _snake_case : Optional[Any] = model[0].eval() # set weights for wav2vec2 encoder _snake_case : Any = WavaVecaModel(snake_case__ ) _snake_case : Optional[Any] = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) _snake_case : Optional[Any] = SpeechaTextaForCausalLM(snake_case__ ) _snake_case , _snake_case : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) _snake_case : Any = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) _snake_case : Any = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) _snake_case : Any = False # add projection layer _snake_case : int = nn.Parameter(projection_layer.weight ) _snake_case : Any = nn.Parameter(projection_layer.bias ) _snake_case : Any = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) _snake_case : Dict = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) _snake_case : str = hf_wavavec.config.to_dict() _snake_case : List[str] = tokenizer.pad_token_id _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Union[str, Any] = tokenizer.eos_token_id _snake_case : Optional[Any] = """speech_to_text_2""" _snake_case : Optional[int] = """wav2vec2""" _snake_case : Tuple = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_02_24, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') A_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
64
1
"""simple docstring""" import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' super().tearDown() gc.collect() def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case , _snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( """stabilityai/stable-diffusion-2""", revision="""bf16""", dtype=jnp.bfloataa, ) _snake_case : str = """A painting of a squirrel eating a burger""" _snake_case : Union[str, Any] = jax.device_count() _snake_case : str = num_samples * [prompt] _snake_case : Union[str, Any] = sd_pipe.prepare_inputs(a_ ) _snake_case : List[Any] = replicate(a_ ) _snake_case : str = shard(a_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Tuple = jax.random.split(a_, jax.device_count() ) _snake_case : Union[str, Any] = sd_pipe(a_, a_, a_, num_inference_steps=25, jit=a_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : Tuple = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[Any] = images[0, 253:256, 253:256, -1] _snake_case : Optional[Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Any = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : str = """stabilityai/stable-diffusion-2""" _snake_case , _snake_case : Dict = FlaxDPMSolverMultistepScheduler.from_pretrained(a_, subfolder="""scheduler""" ) _snake_case , _snake_case : Any = FlaxStableDiffusionPipeline.from_pretrained( a_, scheduler=a_, revision="""bf16""", dtype=jnp.bfloataa, ) _snake_case : Tuple = scheduler_params _snake_case : Dict = """A painting of a squirrel eating a burger""" _snake_case : str = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : str = sd_pipe.prepare_inputs(a_ ) _snake_case : List[str] = replicate(a_ ) _snake_case : Tuple = shard(a_ ) _snake_case : Union[str, Any] = jax.random.PRNGKey(0 ) _snake_case : Dict = jax.random.split(a_, jax.device_count() ) _snake_case : str = sd_pipe(a_, a_, a_, num_inference_steps=25, jit=a_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : Tuple = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[Any] = images[0, 253:256, 253:256, -1] _snake_case : Optional[int] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
64
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''weiweishi/roc-bert-base-zh''': '''https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json''', } class lowercase( __a ): '''simple docstring''' lowercase__ = "roc_bert" def __init__( self: Tuple, a_: Any=30_522, a_: Any=768, a_: int=12, a_: str=12, a_: Tuple=3_072, a_: List[str]="gelu", a_: Union[str, Any]=0.1, a_: str=0.1, a_: Optional[Any]=512, a_: str=2, a_: Any=0.02, a_: Any=1E-12, a_: Optional[Any]=True, a_: int=0, a_: str="absolute", a_: Any=None, a_: List[Any]=True, a_: Optional[Any]=True, a_: str=768, a_: Union[str, Any]=910, a_: Tuple=512, a_: List[str]=24_858, a_: Optional[int]=True, **a_: List[str], ): '''simple docstring''' _snake_case : str = vocab_size _snake_case : Union[str, Any] = max_position_embeddings _snake_case : str = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : Tuple = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : Optional[Any] = hidden_act _snake_case : Dict = hidden_dropout_prob _snake_case : Optional[Any] = attention_probs_dropout_prob _snake_case : Dict = initializer_range _snake_case : Union[str, Any] = type_vocab_size _snake_case : List[str] = layer_norm_eps _snake_case : List[str] = use_cache _snake_case : List[Any] = enable_pronunciation _snake_case : Union[str, Any] = enable_shape _snake_case : Optional[int] = pronunciation_embed_dim _snake_case : int = pronunciation_vocab_size _snake_case : int = shape_embed_dim _snake_case : Dict = shape_vocab_size _snake_case : Optional[int] = concat_input _snake_case : Tuple = position_embedding_type _snake_case : str = classifier_dropout super().__init__(pad_token_id=a_, **a_ )
64
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
1
"""simple docstring""" from pathlib import Path import fire def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Union[str, Any] = Path(snake_case__ ) _snake_case : int = Path(snake_case__ ) dest_dir.mkdir(exist_ok=snake_case__ ) for path in src_dir.iterdir(): _snake_case : List[Any] = [x.rstrip() for x in list(path.open().readlines() )][:n] _snake_case : Tuple = dest_dir.joinpath(path.name ) print(snake_case__ ) dest_path.open("""w""" ).write("""\n""".join(snake_case__ ) ) if __name__ == "__main__": fire.Fire(minify)
64
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''tokenizer_file''': { '''EleutherAI/gpt-neox-20b''': '''https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json''', }, } A_ = { '''gpt-neox-20b''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] def __init__( self: Dict, a_: Dict=None, a_: Tuple=None, a_: Dict=None, a_: str="<|endoftext|>", a_: Union[str, Any]="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: Optional[int]=False, **a_: str, ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) _snake_case : int = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Optional[int] = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Any = add_prefix_space _snake_case : List[str] = pre_tok_class(**a_ ) _snake_case : Tuple = add_prefix_space def UpperCamelCase_ ( self: List[Any], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Dict = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: List[Any], a_: "Conversation" ): '''simple docstring''' _snake_case : List[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(a_, add_special_tokens=a_ ) + [self.eos_token_id] ) if len(a_ ) > self.model_max_length: _snake_case : int = input_ids[-self.model_max_length :] return input_ids
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
1
"""simple docstring""" import csv import tweepy # Twitter API credentials A_ = '''''' A_ = '''''' A_ = '''''' A_ = '''''' def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = tweepy.OAuthHandler(snake_case__ , snake_case__ ) auth.set_access_token(snake_case__ , snake_case__ ) _snake_case : Dict = tweepy.API(snake_case__ ) # initialize a list to hold all the tweepy Tweets _snake_case : str = [] # make initial request for most recent tweets (200 is the maximum allowed count) _snake_case : str = api.user_timeline(screen_name=snake_case__ , count=2_00 ) # save most recent tweets alltweets.extend(snake_case__ ) # save the id of the oldest tweet less one _snake_case : List[str] = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(snake_case__ ) > 0: print(F"getting tweets before {oldest}" ) # all subsequent requests use the max_id param to prevent duplicates _snake_case : int = api.user_timeline( screen_name=snake_case__ , count=2_00 , max_id=snake_case__ ) # save most recent tweets alltweets.extend(snake_case__ ) # update the id of the oldest tweet less one _snake_case : List[str] = alltweets[-1].id - 1 print(F"...{len(snake_case__ )} tweets downloaded so far" ) # transform the tweepy tweets into a 2D array that will populate the csv _snake_case : Tuple = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(F"new_{screen_name}_tweets.csv" , """w""" ) as f: _snake_case : Any = csv.writer(snake_case__ ) writer.writerow(["""id""", """created_at""", """text"""] ) writer.writerows(snake_case__ ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
64
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = CodeGenTokenizer def __init__( self: Union[str, Any], a_: List[Any]=None, a_: str=None, a_: str=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: str="<|endoftext|>", a_: List[Any]=False, **a_: List[str], ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) if kwargs.pop("""add_bos_token""", a_ ): _snake_case : str = kwargs.pop("""name_or_path""", """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" f"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" f"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) _snake_case : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Dict = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Dict = add_prefix_space _snake_case : str = pre_tok_class(**a_ ) _snake_case : List[Any] = add_prefix_space def UpperCamelCase_ ( self: Any, *a_: Any, **a_: int ): '''simple docstring''' _snake_case : Optional[int] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Any, **a_: List[str] ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: str, a_: Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"], a_: bool = False, a_: bool = None, a_: Optional[List[str]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = super().decode( token_ids=a_, skip_special_tokens=a_, clean_up_tokenization_spaces=a_, **a_, ) if truncate_before_pattern is not None and len(a_ ) > 0: _snake_case : List[str] = self.truncate(a_, a_ ) return decoded_text def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Optional[Any] ): '''simple docstring''' def find_re(a_: Dict, a_: str, a_: Union[str, Any] ): _snake_case : Any = pattern.search(a_, a_ ) return m.start() if m else -1 _snake_case : Tuple = [re.compile(a_, re.MULTILINE ) for pattern in truncate_before_pattern] _snake_case : List[Any] = list(re.finditer("""^print""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : int = completion[: prints[1].start()] _snake_case : List[str] = list(re.finditer("""^def""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : List[Any] = completion[: defs[1].start()] _snake_case : int = 0 _snake_case : List[Any] = [ pos for pos in [find_re(a_, a_, a_ ) for terminal in terminals] if pos != -1 ] if len(a_ ) > 0: return completion[: min(a_ )] else: return completion
64
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
64
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
1
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True} ) lowercase__ = Features({"text": Value("string" )} ) lowercase__ = Features({} ) lowercase__ = "text" @property def UpperCamelCase_ ( self: Any ): '''simple docstring''' return {self.text_column: "text"}
64
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str]=False ): """simple docstring""" _snake_case : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def UpperCAmelCase__ (snake_case__ : Dict , snake_case__ : Dict , snake_case__ : List[str]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : List[Any] = """""" else: _snake_case : List[Any] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) _snake_case : Optional[Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _snake_case : Union[str, Any] = in_proj_bias[: config.hidden_size] _snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : List[str] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Tuple = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" _snake_case : List[str] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : int ): """simple docstring""" _snake_case : Optional[Any] = dct.pop(snake_case__ ) _snake_case : Union[str, Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : str = ViTMSNConfig() _snake_case : Any = 10_00 _snake_case : Tuple = """datasets/huggingface/label-files""" _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : int = json.load(open(hf_hub_download(snake_case__ , snake_case__ ) , """r""" ) ) _snake_case : Any = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[Any] = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _snake_case : Tuple = 3_84 _snake_case : Dict = 15_36 _snake_case : Tuple = 6 elif "l16" in checkpoint_url: _snake_case : Any = 10_24 _snake_case : int = 40_96 _snake_case : str = 24 _snake_case : Optional[int] = 16 _snake_case : List[Any] = 0.1 elif "b4" in checkpoint_url: _snake_case : Tuple = 4 elif "l7" in checkpoint_url: _snake_case : int = 7 _snake_case : Dict = 10_24 _snake_case : Optional[Any] = 40_96 _snake_case : Any = 24 _snake_case : Union[str, Any] = 16 _snake_case : Optional[int] = 0.1 _snake_case : int = ViTMSNModel(snake_case__ ) _snake_case : Optional[int] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" )["""target_encoder"""] _snake_case : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(snake_case__ ) _snake_case : List[str] = create_rename_keys(snake_case__ , base_model=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) read_in_q_k_v(snake_case__ , snake_case__ , base_model=snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() _snake_case : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Tuple = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) _snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=snake_case__ , image_std=snake_case__ ) _snake_case : Any = image_processor(images=snake_case__ , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) _snake_case : int = model(**snake_case__ ) _snake_case : List[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _snake_case : Optional[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: _snake_case : str = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: _snake_case : Optional[int] = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: _snake_case : List[Any] = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: _snake_case : Optional[int] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , snake_case__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if not isinstance(snake_case__ , snake_case__ ): _snake_case : Tuple = F"Input value of [number={number}] must be an integer" raise TypeError(snake_case__ ) if number < 0: return False _snake_case : Optional[Any] = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class lowercase: '''simple docstring''' lowercase__ = PegasusConfig lowercase__ = {} lowercase__ = "gelu" def __init__( self: List[str], a_: Any, a_: Union[str, Any]=13, a_: Tuple=7, a_: int=True, a_: List[Any]=False, a_: int=99, a_: str=32, a_: str=2, a_: List[str]=4, a_: Union[str, Any]=37, a_: List[Any]=0.1, a_: List[str]=0.1, a_: List[Any]=40, a_: Optional[int]=2, a_: Dict=1, a_: Dict=0, ): '''simple docstring''' _snake_case : Dict = parent _snake_case : List[str] = batch_size _snake_case : Optional[int] = seq_length _snake_case : List[str] = is_training _snake_case : Any = use_labels _snake_case : List[str] = vocab_size _snake_case : Optional[int] = hidden_size _snake_case : Tuple = num_hidden_layers _snake_case : Optional[Any] = num_attention_heads _snake_case : str = intermediate_size _snake_case : Optional[Any] = hidden_dropout_prob _snake_case : Optional[Any] = attention_probs_dropout_prob _snake_case : List[Any] = max_position_embeddings _snake_case : Optional[Any] = eos_token_id _snake_case : Dict = pad_token_id _snake_case : List[Any] = bos_token_id def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ) _snake_case : Optional[Any] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ), 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor], axis=1 ) _snake_case : int = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) _snake_case : Any = self.config_cls( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_ids=[2], bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.pad_token_id, **self.config_updates, ) _snake_case : Optional[int] = prepare_pegasus_inputs_dict(a_, a_, a_ ) return config, inputs_dict def UpperCamelCase_ ( self: List[str], a_: Union[str, Any], a_: Optional[Any] ): '''simple docstring''' _snake_case : Any = TFPegasusModel(config=a_ ).get_decoder() _snake_case : Dict = inputs_dict["""input_ids"""] _snake_case : int = input_ids[:1, :] _snake_case : Optional[Any] = inputs_dict["""attention_mask"""][:1, :] _snake_case : List[Any] = inputs_dict["""head_mask"""] _snake_case : Any = 1 # first forward pass _snake_case : Dict = model(a_, attention_mask=a_, head_mask=a_, use_cache=a_ ) _snake_case , _snake_case : int = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Optional[Any] = ids_tensor((self.batch_size, 3), config.vocab_size ) _snake_case : int = tf.cast(ids_tensor((self.batch_size, 3), 2 ), tf.inta ) # append to next input_ids and _snake_case : Dict = tf.concat([input_ids, next_tokens], axis=-1 ) _snake_case : Dict = tf.concat([attention_mask, next_attn_mask], axis=-1 ) _snake_case : Optional[int] = model(a_, attention_mask=a_ )[0] _snake_case : Optional[int] = model(a_, attention_mask=a_, past_key_values=a_ )[0] self.parent.assertEqual(next_tokens.shape[1], output_from_past.shape[1] ) # select random slice _snake_case : Any = int(ids_tensor((1,), output_from_past.shape[-1] ) ) _snake_case : Optional[int] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : int = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(a_, a_, rtol=1E-3 ) def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : List[Any]=None , snake_case__ : Optional[int]=None , snake_case__ : Dict=None , snake_case__ : Tuple=None , snake_case__ : Tuple=None , ): """simple docstring""" if attention_mask is None: _snake_case : str = tf.cast(tf.math.not_equal(snake_case__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : int = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : List[Any] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : int = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case : Union[str, Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowercase( __a , __a , unittest.TestCase ): '''simple docstring''' lowercase__ = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () lowercase__ = (TFPegasusForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { "conversational": TFPegasusForConditionalGeneration, "feature-extraction": TFPegasusModel, "summarization": TFPegasusForConditionalGeneration, "text2text-generation": TFPegasusForConditionalGeneration, "translation": TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : str = TFPegasusModelTester(self ) _snake_case : Any = ConfigTester(self, config_class=a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*a_ ) @require_sentencepiece @require_tokenizers @require_tf class lowercase( unittest.TestCase ): '''simple docstring''' lowercase__ = [ " PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.", " The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" ", ] lowercase__ = [ "California's largest electricity provider has cut power to hundreds of thousands of customers in an effort to" " reduce the risk of wildfires.", "N-Dubz have revealed they\'re \"grateful\" to have been nominated for four Mobo Awards.", ] # differs slightly from pytorch, likely due to numerical differences in linear layers lowercase__ = "google/pegasus-xsum" @cached_property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def UpperCamelCase_ ( self: Optional[Any], **a_: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = self.translate_src_text(**a_ ) assert self.expected_text == generated_words def UpperCamelCase_ ( self: Tuple, **a_: int ): '''simple docstring''' _snake_case : Tuple = self.tokenizer(self.src_text, **a_, padding=a_, return_tensors="""tf""" ) _snake_case : Optional[int] = self.model.generate( model_inputs.input_ids, attention_mask=model_inputs.attention_mask, num_beams=2, use_cache=a_, ) _snake_case : List[Any] = self.tokenizer.batch_decode(generated_ids.numpy(), skip_special_tokens=a_ ) return generated_words @slow def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' self._assert_generated_batch_equal_expected()
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" stooge(snake_case__ , 0 , len(snake_case__ ) - 1 ) return arr def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any , snake_case__ : int ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) # Recursively sort last 2/3 elements stooge(snake_case__ , i + t , (snake_case__) ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list[list] ): """simple docstring""" _snake_case : Tuple = current_set.copy() for row_index, row in enumerate(snake_case__ ): _snake_case : Tuple = row[0] for column_index, column in enumerate(snake_case__ ): if magnitude == 0: _snake_case : List[Any] = column continue _snake_case : Tuple = column / magnitude # Subtract to cancel term _snake_case : Optional[Any] = current_set[0] _snake_case : List[str] = [first_row] _snake_case : str = current_set[1::] for row in current_set: _snake_case : int = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(snake_case__ ) continue for column_index in range(len(snake_case__ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(snake_case__ ) # Create next recursion iteration set if len(final_set[0] ) != 3: _snake_case : Union[str, Any] = final_set[0] _snake_case : int = [] _snake_case : str = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) _snake_case : Optional[Any] = simplify(snake_case__ ) for i in range(len(snake_case__ ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , snake_case__ ) _snake_case : List[str] = resultant return final_set def UpperCAmelCase__ (snake_case__ : list[list] ): """simple docstring""" if len(snake_case__ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) _snake_case : Union[str, Any] = len(snake_case__ ) + 1 if any(len(snake_case__ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(snake_case__ , (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(snake_case__ ) == 1: return [equations[0][-1] / equations[0][0]] _snake_case : Dict = equations.copy() if any(0 in row for row in data_set ): _snake_case : Any = data_set.copy() _snake_case : int = [] for row_index, row in enumerate(snake_case__ ): if 0 not in row: _snake_case : Optional[int] = data_set.pop(snake_case__ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0 , snake_case__ ) _snake_case : str = data_set.copy() _snake_case : Union[str, Any] = simplify(snake_case__ ) _snake_case : int = simplified[::-1] _snake_case : list = [] for row in simplified: _snake_case : int = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue _snake_case : Dict = row.copy()[: len(snake_case__ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(snake_case__ ) == 0: solutions.append(0 ) continue _snake_case : Optional[Any] = temp_row[1::] _snake_case : Dict = temp_row[::-1] for column_index, column in enumerate(snake_case__ ): current_solution -= column * solutions[column_index] solutions.append(snake_case__ ) _snake_case : str = [] for item in solutions: final.append(float(round(snake_case__ , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() A_ = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
64
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowercase( metaclass=__a ): '''simple docstring''' lowercase__ = ["note_seq"] def __init__( self: Dict, *a_: Union[str, Any], **a_: List[str] ): '''simple docstring''' requires_backends(self, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Optional[int], *a_: Any, **a_: Optional[Any] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] ) @classmethod def UpperCamelCase_ ( cls: Tuple, *a_: Optional[Any], **a_: List[str] ): '''simple docstring''' requires_backends(cls, ["""note_seq"""] )
64
1
"""simple docstring""" from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = [] for part_id in partition_order: _snake_case : str = df.where(F"SPARK_PARTITION_ID() = {part_id}" ).collect() for row_idx, row in enumerate(snake_case__ ): expected_row_ids_and_row_dicts.append((F"{part_id}_{row_idx}", row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : Optional[Any] = spark.range(1_00 ).repartition(1 ) _snake_case : Tuple = Spark(snake_case__ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : Tuple = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : int = spark.range(10 ).repartition(2 ) _snake_case : Tuple = [1, 0] _snake_case : Dict = _generate_iterable_examples(snake_case__ , snake_case__ ) # Reverse the partitions. _snake_case : Optional[int] = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case__ , snake_case__ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): _snake_case , _snake_case : Optional[Any] = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : Any = spark.range(10 ).repartition(1 ) _snake_case : List[str] = SparkExamplesIterable(snake_case__ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(snake_case__ ): assert row_id == F"0_{i}" assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : Tuple = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : int = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch("""numpy.random.Generator""" ) as generator_mock: _snake_case : Tuple = lambda snake_case__ : x.reverse() _snake_case : List[str] = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case__ , [2, 1, 0] ) _snake_case : int = SparkExamplesIterable(snake_case__ ).shuffle_data_sources(snake_case__ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(snake_case__ ): _snake_case , _snake_case : List[Any] = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : Dict = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 _snake_case : List[Any] = SparkExamplesIterable(snake_case__ ).shard_data_sources(worker_id=0 , num_workers=2 ) assert shard_it_a.n_shards == 2 _snake_case : Any = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case__ , [0, 2] ) for i, (row_id, row_dict) in enumerate(snake_case__ ): _snake_case , _snake_case : List[str] = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 _snake_case : List[str] = SparkExamplesIterable(snake_case__ ).shard_data_sources(worker_id=1 , num_workers=2 ) assert shard_it_a.n_shards == 2 _snake_case : Union[str, Any] = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case__ , [1, 3] ) for i, (row_id, row_dict) in enumerate(snake_case__ ): _snake_case , _snake_case : Any = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() _snake_case : Dict = spark.range(1_00 ).repartition(1 ) _snake_case : List[Any] = Spark(snake_case__ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 1_00
64
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: List[str] ): '''simple docstring''' _snake_case : int = data _snake_case : Dict = [0X67452301, 0Xefcdab89, 0X98badcfe, 0X10325476, 0Xc3d2e1f0] @staticmethod def UpperCamelCase_ ( a_: Optional[Any], a_: Dict ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0Xffffffff def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Union[str, Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _snake_case : Optional[int] = self.data + padding + struct.pack(""">Q""", 8 * len(self.data ) ) return padded_data def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0, len(self.padded_data ), 64 ) ] def UpperCamelCase_ ( self: Optional[Any], a_: List[Any] ): '''simple docstring''' _snake_case : List[str] = list(struct.unpack(""">16L""", a_ ) ) + [0] * 64 for i in range(16, 80 ): _snake_case : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]), 1 ) return w def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = self.padding() _snake_case : str = self.split_blocks() for block in self.blocks: _snake_case : Any = self.expand_block(a_ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = self.h for i in range(0, 80 ): if 0 <= i < 20: _snake_case : int = (b & c) | ((~b) & d) _snake_case : str = 0X5a827999 elif 20 <= i < 40: _snake_case : Optional[int] = b ^ c ^ d _snake_case : str = 0X6ed9eba1 elif 40 <= i < 60: _snake_case : List[Any] = (b & c) | (b & d) | (c & d) _snake_case : List[Any] = 0X8f1bbcdc elif 60 <= i < 80: _snake_case : List[Any] = b ^ c ^ d _snake_case : int = 0Xca62c1d6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : Optional[int] = ( self.rotate(a_, 5 ) + f + e + k + expanded_block[i] & 0Xffffffff, a, self.rotate(a_, 30 ), c, d, ) _snake_case : Union[str, Any] = ( self.h[0] + a & 0Xffffffff, self.h[1] + b & 0Xffffffff, self.h[2] + c & 0Xffffffff, self.h[3] + d & 0Xffffffff, self.h[4] + e & 0Xffffffff, ) return ("{:08x}" * 5).format(*self.h ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = B"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[Any] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) _snake_case : Union[str, Any] = parser.parse_args() _snake_case : List[Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: _snake_case : str = f.read() else: _snake_case : int = bytes(snake_case__ , """utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
64
1
"""simple docstring""" from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class lowercase( nn.Module ): '''simple docstring''' def __init__( self: Any, a_: int = 16, a_: int = 88, a_: Optional[int] = None, a_: int = 1, a_: float = 0.0, a_: int = 32, a_: Optional[int] = None, a_: bool = False, a_: Optional[int] = None, a_: Optional[int] = None, a_: str = "geglu", a_: Optional[int] = None, ): '''simple docstring''' super().__init__() _snake_case : Optional[int] = nn.ModuleList( [ TransformeraDModel( num_attention_heads=a_, attention_head_dim=a_, in_channels=a_, num_layers=a_, dropout=a_, norm_num_groups=a_, cross_attention_dim=a_, attention_bias=a_, sample_size=a_, num_vector_embeds=a_, activation_fn=a_, num_embeds_ada_norm=a_, ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference _snake_case : Optional[Any] = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` _snake_case : str = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` _snake_case : Any = [1, 0] def UpperCamelCase_ ( self: Any, a_: Optional[int], a_: Optional[Any], a_: int=None, a_: str=None, a_: List[str]=None, a_: bool = True, ): '''simple docstring''' _snake_case : List[str] = hidden_states _snake_case : Tuple = [] _snake_case : Optional[int] = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens _snake_case : str = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] _snake_case : Any = self.transformer_index_for_condition[i] _snake_case : int = self.transformers[transformer_index]( a_, encoder_hidden_states=a_, timestep=a_, cross_attention_kwargs=a_, return_dict=a_, )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] _snake_case : List[Any] = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) _snake_case : List[Any] = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=a_ )
64
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
1
"""simple docstring""" import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class lowercase: '''simple docstring''' def __init__( self: List[Any], a_: Optional[int], a_: Optional[int]=13, a_: List[str]=7, a_: List[str]=True, a_: Dict=True, a_: Tuple=True, a_: List[Any]=True, a_: List[Any]=99, a_: Any=32, a_: Optional[int]=5, a_: int=4, a_: List[Any]=4, a_: Tuple="gelu", a_: Optional[int]=0.0, a_: Tuple=0.1, a_: List[Any]=True, a_: int=512, a_: Optional[Any]=16, a_: Union[str, Any]=2, a_: Dict=0.02, a_: Optional[Any]=3, a_: Dict=4, a_: Dict=None, ): '''simple docstring''' _snake_case : Optional[int] = parent _snake_case : Optional[int] = batch_size _snake_case : str = seq_length _snake_case : Dict = is_training _snake_case : Optional[int] = use_input_mask _snake_case : Dict = use_token_type_ids _snake_case : str = use_labels _snake_case : Tuple = vocab_size _snake_case : int = hidden_size _snake_case : List[Any] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : str = intermediate_multiple_size _snake_case : List[str] = hidden_act _snake_case : Dict = hidden_dropout _snake_case : List[Any] = attention_dropout _snake_case : Optional[Any] = weight_tying _snake_case : int = max_position_embeddings _snake_case : List[Any] = type_vocab_size _snake_case : int = type_sequence_label_size _snake_case : Any = initializer_range _snake_case : str = num_labels _snake_case : Tuple = num_choices _snake_case : Optional[Any] = scope def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) _snake_case : str = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : Optional[int] = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) _snake_case : List[str] = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self: Any ): '''simple docstring''' return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_multiple_size=self.intermediate_multiple_size, hidden_act=self.hidden_act, hidden_dropout=self.hidden_dropout, attention_dropout=self.attention_dropout, weight_tying=self.weight_tying, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, is_decoder=a_, initializer_range=self.initializer_range, ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case , _snake_case , _snake_case , _snake_case : List[str] = self.prepare_config_and_inputs() _snake_case : Optional[Any] = True return config, input_ids, input_mask, token_labels def UpperCamelCase_ ( self: List[str], a_: Optional[int], a_: Optional[int], a_: Optional[int] ): '''simple docstring''' _snake_case : List[str] = GPTNeoXJapaneseModel(config=a_ ) model.to(a_ ) model.eval() _snake_case : List[str] = model(a_, attention_mask=a_ ) _snake_case : Union[str, Any] = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self: str, a_: Tuple, a_: Optional[int], a_: str ): '''simple docstring''' _snake_case : Optional[int] = True _snake_case : int = GPTNeoXJapaneseModel(a_ ) model.to(a_ ) model.eval() _snake_case : Optional[int] = model(a_, attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self: List[str], a_: int, a_: Union[str, Any], a_: Any, a_: List[str] ): '''simple docstring''' _snake_case : List[str] = GPTNeoXJapaneseForCausalLM(config=a_ ) model.to(a_ ) model.eval() _snake_case : Optional[int] = model(a_, attention_mask=a_, labels=a_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self: int, a_: Optional[int], a_: str, a_: Optional[int] ): '''simple docstring''' _snake_case : Tuple = True _snake_case : Optional[Any] = GPTNeoXJapaneseForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass _snake_case : List[Any] = model(a_, attention_mask=a_, use_cache=a_ ) _snake_case : List[str] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _snake_case : List[Any] = ids_tensor((self.batch_size, 3), config.vocab_size ) _snake_case : Dict = ids_tensor((self.batch_size, 3), vocab_size=2 ) # append to next input_ids and _snake_case : Dict = torch.cat([input_ids, next_tokens], dim=-1 ) _snake_case : Optional[int] = torch.cat([input_mask, next_mask], dim=-1 ) _snake_case : Union[str, Any] = model(a_, attention_mask=a_, output_hidden_states=a_ ) _snake_case : Optional[Any] = output_from_no_past["""hidden_states"""][0] _snake_case : Optional[int] = model( a_, attention_mask=a_, past_key_values=a_, output_hidden_states=a_, )["""hidden_states"""][0] # select random slice _snake_case : Tuple = ids_tensor((1,), output_from_past.shape[-1] ).item() _snake_case : str = output_from_no_past[:, -3:, random_slice_idx].detach() _snake_case : str = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_, a_, atol=1E-3 ) ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : Tuple = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case , _snake_case : Any = config_and_inputs _snake_case : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowercase( __a , __a , unittest.TestCase ): '''simple docstring''' lowercase__ = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () lowercase__ = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () lowercase__ = ( {"feature-extraction": GPTNeoXJapaneseModel, "text-generation": GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : List[str] = GPTNeoXJapaneseModelTester(self ) _snake_case : Union[str, Any] = ConfigTester(self, config_class=a_, hidden_size=37 ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case , _snake_case , _snake_case , _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(a_, a_, a_ ) def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case , _snake_case , _snake_case , _snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(a_, a_, a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case , _snake_case , _snake_case , _snake_case : int = self.model_tester.prepare_config_and_inputs_for_decoder() _snake_case : Union[str, Any] = None self.model_tester.create_and_check_model_as_decoder(a_, a_, a_ ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case , _snake_case , _snake_case , _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(a_, a_, a_ ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*a_ ) @slow def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Optional[int] = """abeja/gpt-neox-japanese-2.7b""" _snake_case : List[Any] = ["""データサイエンティストとは、""", """100年後に必要とされる会社は、""", """フルリモートの環境で働くために必要なことは、""", """国境の長いトンネルを抜けると""", """美味しい日本食といえば、"""] _snake_case : int = [ """データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。""", """100年後に必要とされる会社は、「人」が中心の会社です。""", """フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。""", """国境の長いトンネルを抜けると、そこは雪国だった。""", """美味しい日本食といえば、やっぱりお寿司ですよね。""", ] _snake_case : Union[str, Any] = GPTNeoXJapaneseTokenizer.from_pretrained(a_ ) _snake_case : str = GPTNeoXJapaneseForCausalLM.from_pretrained(a_ ) _snake_case : Tuple = [] for prompt in prompts: _snake_case : List[Any] = tokenizer(a_, return_tensors="""pt""" ).input_ids _snake_case : Tuple = model.generate(a_, max_length=50 ) _snake_case : str = tokenizer.batch_decode(a_, skip_special_tokens=a_ ) predicted_outputs += generated_string self.assertListEqual(a_, a_ )
64
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
1
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path A_ = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) A_ = [ord(letter) for letter in string.ascii_lowercase] A_ = {ord(char) for char in VALID_CHARS} A_ = ["the", "be", "to", "of", "and", "in", "that", "have"] def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : tuple[int, ...] ): """simple docstring""" _snake_case : str = "" _snake_case : int _snake_case : int _snake_case : int for keychar, cipherchar in zip(cycle(snake_case__ ) , snake_case__ ): _snake_case : Dict = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(snake_case__ ) return decoded def UpperCAmelCase__ (snake_case__ : list[int] ): """simple docstring""" _snake_case : list[str] = [] for key in product(snake_case__ , repeat=3 ): _snake_case : List[str] = try_key(snake_case__ , snake_case__ ) if encoded is not None: possibles.append(snake_case__ ) return possibles def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : str ): """simple docstring""" return [possible for possible in possibles if common_word in possible.lower()] def UpperCAmelCase__ (snake_case__ : str = "p059_cipher.txt" ): """simple docstring""" _snake_case : list[int] _snake_case : list[str] _snake_case : str _snake_case : str _snake_case : str = Path(snake_case__ ).parent.joinpath(snake_case__ ).read_text(encoding="""utf-8""" ) _snake_case : Union[str, Any] = [int(snake_case__ ) for number in data.strip().split(""",""" )] _snake_case : str = filter_valid_chars(snake_case__ ) for common_word in COMMON_WORDS: _snake_case : int = filter_common_word(snake_case__ , snake_case__ ) if len(snake_case__ ) == 1: break _snake_case : List[str] = possibles[0] return sum(ord(snake_case__ ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : list ): """simple docstring""" if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] _snake_case : List[Any] = [] def generate(snake_case__ : int , snake_case__ : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _snake_case , _snake_case : Optional[Any] = arr[k - 1], arr[i] else: # k is odd _snake_case , _snake_case : List[str] = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
64
1
"""simple docstring""" import baseaa def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" return baseaa.baaencode(string.encode("""utf-8""" ) ) def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" return baseaa.baadecode(snake_case__ ).decode("""utf-8""" ) if __name__ == "__main__": A_ = '''Hello World!''' A_ = baseaa_encode(test) print(encoded) A_ = baseaa_decode(encoded) print(decoded)
64
"""simple docstring""" from math import factorial A_ = {str(d): factorial(d) for d in range(10)} def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" return sum(DIGIT_FACTORIAL[d] for d in str(snake_case__ ) ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : List[str] = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , snake_case__ ) if sum_of_digit_factorial(snake_case__ ) == i ) if __name__ == "__main__": print(F'''{solution() = }''')
64
1
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class lowercase( __a ): '''simple docstring''' lowercase__ = ( "This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image." "It takes two arguments named `image` which should be the original image, and `label` which should be a text " "describing the elements what should be identified in the segmentation mask. The tool returns the mask." ) lowercase__ = "CIDAS/clipseg-rd64-refined" lowercase__ = "image_segmenter" lowercase__ = CLIPSegForImageSegmentation lowercase__ = ["image", "text"] lowercase__ = ["image"] def __init__( self: Dict, *a_: str, **a_: Tuple ): '''simple docstring''' requires_backends(self, ["""vision"""] ) super().__init__(*a_, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: "Image", a_: str ): '''simple docstring''' return self.pre_processor(text=[label], images=[image], padding=a_, return_tensors="""pt""" ) def UpperCamelCase_ ( self: int, a_: Tuple ): '''simple docstring''' with torch.no_grad(): _snake_case : Dict = self.model(**a_ ).logits return logits def UpperCamelCase_ ( self: Any, a_: Any ): '''simple docstring''' _snake_case : Optional[Any] = outputs.cpu().detach().numpy() _snake_case : Tuple = 0 _snake_case : Optional[Any] = 1 return Image.fromarray((array * 255).astype(np.uinta ) )
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[int] , snake_case__ : int ): """simple docstring""" if len(snake_case__ ) < k or k < 0: raise ValueError("""Invalid Input""" ) _snake_case : Optional[int] = sum(array[:k] ) for i in range(len(snake_case__ ) - k ): _snake_case : Optional[Any] = current_sum - array[i] + array[i + k] _snake_case : List[str] = max(snake_case__ , snake_case__ ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-10_00, 10_00) for i in range(1_00)] A_ = randint(0, 1_10) print(F'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
64
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) A_ = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.encoder.norm.weight''', '''encoder.layernorm.weight'''), ('''transformer.encoder.norm.bias''', '''encoder.layernorm.bias'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ] ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Tuple , snake_case__ : List[Any] ): """simple docstring""" _snake_case : List[Any] = state_dict.pop(snake_case__ ) _snake_case : Optional[Any] = val def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case : Dict = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) _snake_case : List[Any] = value else: _snake_case : Optional[Any] = value return new_state_dict def UpperCAmelCase__ (snake_case__ : Tuple ): """simple docstring""" _snake_case : Tuple = """""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case : Dict = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) _snake_case : Optional[int] = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[:2_56, :] _snake_case : Union[str, Any] = in_proj_bias[:2_56] _snake_case : List[str] = in_proj_weight[2_56:5_12, :] _snake_case : Optional[Any] = in_proj_bias[2_56:5_12] _snake_case : int = in_proj_weight[-2_56:, :] _snake_case : Any = in_proj_bias[-2_56:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case : int = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) _snake_case : Optional[int] = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : str = in_proj_weight[:2_56, :] _snake_case : Optional[int] = in_proj_bias[:2_56] _snake_case : Optional[Any] = in_proj_weight[2_56:5_12, :] _snake_case : Optional[int] = in_proj_bias[2_56:5_12] _snake_case : Dict = in_proj_weight[-2_56:, :] _snake_case : List[str] = in_proj_bias[-2_56:] # read in weights + bias of input projection layer of cross-attention _snake_case : List[Any] = state_dict.pop( F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight" ) _snake_case : Any = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias" ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case : List[Any] = in_proj_weight_cross_attn[:2_56, :] _snake_case : int = in_proj_bias_cross_attn[:2_56] _snake_case : Union[str, Any] = in_proj_weight_cross_attn[2_56:5_12, :] _snake_case : Optional[int] = in_proj_bias_cross_attn[2_56:5_12] _snake_case : str = in_proj_weight_cross_attn[-2_56:, :] _snake_case : Tuple = in_proj_bias_cross_attn[-2_56:] def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : Optional[Any] ): """simple docstring""" _snake_case , _snake_case : int = image.size _snake_case : Optional[Any] = max(snake_case__ , snake_case__ ) _snake_case : Dict = 8_00 if """detection""" in checkpoint_url else 10_00 _snake_case : Dict = target_max_size / current_max_size _snake_case : Any = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def UpperCAmelCase__ (snake_case__ : Any ): """simple docstring""" _snake_case : int = F.to_tensor(snake_case__ ) _snake_case : str = F.normalize(snake_case__ , mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ) return image @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : List[str] ): """simple docstring""" logger.info("""Converting model...""" ) # load original state dict _snake_case : List[Any] = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""" ) # rename keys for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) _snake_case : Any = rename_backbone_keys(snake_case__ ) # query, key and value matrices need special treatment read_in_q_k_v(snake_case__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case : int = """model.""" for key in state_dict.copy().keys(): if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): _snake_case : List[Any] = state_dict.pop(snake_case__ ) _snake_case : Optional[int] = val # create HuggingFace model and load state dict _snake_case : Any = TableTransformerConfig( backbone="""resnet18""" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case : Optional[Any] = 15 _snake_case : Optional[int] = 2 _snake_case : str = {0: """table""", 1: """table rotated"""} _snake_case : str = idalabel _snake_case : List[Any] = {v: k for k, v in idalabel.items()} else: _snake_case : Optional[Any] = 1_25 _snake_case : Optional[int] = 6 _snake_case : Tuple = { 0: """table""", 1: """table column""", 2: """table row""", 3: """table column header""", 4: """table projected row header""", 5: """table spanning cell""", } _snake_case : Dict = idalabel _snake_case : Any = {v: k for k, v in idalabel.items()} _snake_case : List[Any] = DetrImageProcessor( format="""coco_detection""" , max_size=8_00 if """detection""" in checkpoint_url else 10_00 ) _snake_case : str = TableTransformerForObjectDetection(snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() # verify our conversion _snake_case : Union[str, Any] = """example_pdf.png""" if """detection""" in checkpoint_url else """example_table.png""" _snake_case : str = hf_hub_download(repo_id="""nielsr/example-pdf""" , repo_type="""dataset""" , filename=snake_case__ ) _snake_case : Optional[int] = Image.open(snake_case__ ).convert("""RGB""" ) _snake_case : Any = normalize(resize(snake_case__ , snake_case__ ) ).unsqueeze(0 ) _snake_case : Tuple = model(snake_case__ ) if "detection" in checkpoint_url: _snake_case : List[str] = (1, 15, 3) _snake_case : Union[str, Any] = torch.tensor( [[-6.78_97, -16.99_85, 6.79_37], [-8.01_86, -22.21_92, 6.96_77], [-7.31_17, -21.07_08, 7.40_55]] ) _snake_case : str = torch.tensor([[0.48_67, 0.17_67, 0.67_32], [0.67_18, 0.44_79, 0.38_30], [0.47_16, 0.17_60, 0.63_64]] ) else: _snake_case : Tuple = (1, 1_25, 7) _snake_case : Dict = torch.tensor( [[-18.14_30, -8.32_14, 4.82_74], [-18.46_85, -7.13_61, -4.26_67], [-26.36_93, -9.34_29, -4.99_62]] ) _snake_case : Any = torch.tensor([[0.49_83, 0.55_95, 0.94_40], [0.49_16, 0.63_15, 0.59_54], [0.61_08, 0.86_37, 0.11_35]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: # Push model to HF hub logger.info("""Pushing model to the hub...""" ) _snake_case : List[Any] = ( """microsoft/table-transformer-detection""" if """detection""" in checkpoint_url else """microsoft/table-transformer-structure-recognition""" ) model.push_to_hub(snake_case__ ) image_processor.push_to_hub(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_url''', default='''https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth''', type=str, choices=[ '''https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth''', '''https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth''', ], help='''URL of the Table Transformer checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule A_ = {'''processing_wav2vec2_with_lm''': ['''Wav2Vec2ProcessorWithLM''']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
64
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = CodeGenTokenizer lowercase__ = CodeGenTokenizerFast lowercase__ = True lowercase__ = {"add_prefix_space": True} lowercase__ = False def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : Tuple = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] _snake_case : Tuple = dict(zip(a_, range(len(a_ ) ) ) ) _snake_case : str = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _snake_case : List[Any] = {"""unk_token""": """<unk>"""} _snake_case : Optional[int] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(a_ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(a_ ) ) def UpperCamelCase_ ( self: Any, **a_: int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Any, **a_: str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname, **a_ ) def UpperCamelCase_ ( self: Union[str, Any], a_: Dict ): '''simple docstring''' _snake_case : Union[str, Any] = """lower newer""" _snake_case : Tuple = """lower newer""" return input_text, output_text def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = CodeGenTokenizer(self.vocab_file, self.merges_file, **self.special_tokens_map ) _snake_case : Optional[Any] = """lower newer""" _snake_case : Optional[int] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] _snake_case : int = tokenizer.tokenize(a_, add_prefix_space=a_ ) self.assertListEqual(a_, a_ ) _snake_case : str = tokens + [tokenizer.unk_token] _snake_case : Optional[int] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : int = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : Dict = """lower newer""" # Testing tokenization _snake_case : Dict = tokenizer.tokenize(a_, add_prefix_space=a_ ) _snake_case : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids without special tokens _snake_case : Optional[Any] = tokenizer.encode(a_, add_special_tokens=a_, add_prefix_space=a_ ) _snake_case : Tuple = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) # Testing conversion to ids with special tokens _snake_case : Tuple = self.get_rust_tokenizer(add_prefix_space=a_ ) _snake_case : int = tokenizer.encode(a_, add_prefix_space=a_ ) _snake_case : Optional[Any] = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) # Testing the unknown token _snake_case : Tuple = tokens + [rust_tokenizer.unk_token] _snake_case : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ), a_ ) def UpperCamelCase_ ( self: Dict, *a_: Dict, **a_: int ): '''simple docstring''' pass def UpperCamelCase_ ( self: int, a_: List[Any]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) # Simple input _snake_case : Any = """This is a simple input""" _snake_case : Optional[int] = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") _snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Simple input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) # Pair input self.assertRaises(a_, tokenizer_r.encode, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises(a_, tokenizer_r.encode_plus, a_, max_length=a_, padding="""max_length""" ) # Pair input self.assertRaises( a_, tokenizer_r.batch_encode_plus, a_, max_length=a_, padding="""max_length""", ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname, pad_token="""<pad>""" ) # Simple input _snake_case : List[Any] = """This is a simple input""" _snake_case : int = ["""This is a simple input looooooooong""", """This is a simple input"""] _snake_case : Any = ("""This is a simple input""", """This is a pair""") _snake_case : str = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] _snake_case : str = tokenizer.pad_token_id _snake_case : Optional[int] = tokenizer(a_, padding="""max_length""", max_length=30, return_tensors="""np""" ) _snake_case : Dict = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) _snake_case : Tuple = tokenizer(*a_, padding="""max_length""", max_length=60, return_tensors="""np""" ) _snake_case : Optional[Any] = tokenizer(a_, padding=a_, truncate=a_, return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1], 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1], 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1], 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1], 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = """$$$""" _snake_case : List[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname, bos_token=a_, add_bos_token=a_ ) _snake_case : str = """This is a simple input""" _snake_case : int = ["""This is a simple input 1""", """This is a simple input 2"""] _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Tuple = tokenizer(a_ ) _snake_case : Optional[Any] = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0], a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _snake_case : Optional[int] = tokenizer.decode(out_s.input_ids ) _snake_case : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0], a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) _snake_case : Dict = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" _snake_case : Union[str, Any] = """\nif len_a > len_b: result = a\nelse: result = b""" _snake_case : Optional[Any] = tokenizer.encode(a_ ) _snake_case : Dict = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] _snake_case : Optional[Any] = tokenizer.decode(a_, truncate_before_pattern=a_ ) self.assertEqual(a_, a_ ) def UpperCamelCase_ ( self: str ): '''simple docstring''' pass
64
1
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Tuple , snake_case__ : List[Any] ): """simple docstring""" _snake_case : str = WavaVecaForSequenceClassification.from_pretrained(snake_case__ , config=snake_case__ ) _snake_case : List[Any] = downstream_dict["""projector.weight"""] _snake_case : Optional[Any] = downstream_dict["""projector.bias"""] _snake_case : int = downstream_dict["""model.post_net.linear.weight"""] _snake_case : List[str] = downstream_dict["""model.post_net.linear.bias"""] return model def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : Tuple ): """simple docstring""" _snake_case : List[str] = WavaVecaForAudioFrameClassification.from_pretrained(snake_case__ , config=snake_case__ ) _snake_case : Optional[int] = downstream_dict["""model.linear.weight"""] _snake_case : Dict = downstream_dict["""model.linear.bias"""] return model def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = WavaVecaForXVector.from_pretrained(snake_case__ , config=snake_case__ ) _snake_case : Optional[Any] = downstream_dict["""connector.weight"""] _snake_case : Optional[int] = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): _snake_case : Tuple = downstream_dict[ F"model.framelevel_feature_extractor.module.{i}.kernel.weight" ] _snake_case : str = downstream_dict[F"model.framelevel_feature_extractor.module.{i}.kernel.bias"] _snake_case : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] _snake_case : Union[str, Any] = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] _snake_case : List[Any] = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] _snake_case : Dict = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] _snake_case : List[str] = downstream_dict["""objective.W"""] return model @torch.no_grad() def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : Dict , snake_case__ : Dict ): """simple docstring""" _snake_case : List[str] = torch.load(snake_case__ , map_location="""cpu""" ) _snake_case : Union[str, Any] = checkpoint["""Downstream"""] _snake_case : Optional[Any] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( snake_case__ , return_attention_mask=snake_case__ , do_normalize=snake_case__ ) _snake_case : Optional[int] = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): _snake_case : Optional[Any] = convert_classification(snake_case__ , snake_case__ , snake_case__ ) elif arch.endswith("""ForAudioFrameClassification""" ): _snake_case : Dict = convert_diarization(snake_case__ , snake_case__ , snake_case__ ) elif arch.endswith("""ForXVector""" ): _snake_case : List[str] = convert_xvector(snake_case__ , snake_case__ , snake_case__ ) else: raise NotImplementedError(F"S3PRL weights conversion is not supported for {arch}" ) if hf_config.use_weighted_layer_sum: _snake_case : Optional[int] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(snake_case__ ) hf_model.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument( '''--base_model_name''', default=None, type=str, help='''Name of the huggingface pretrained base model.''' ) parser.add_argument('''--config_path''', default=None, type=str, help='''Path to the huggingface classifier config.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to the s3prl checkpoint.''') parser.add_argument('''--model_dump_path''', default=None, type=str, help='''Path to the final converted model.''') A_ = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
64
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ = re.compile(r'''\s+''') def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" return {"hash": hashlib.mda(re.sub(snake_case__ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : Any = [len(snake_case__ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(snake_case__ ), "line_max": max(snake_case__ )} def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : List[Any] ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCAmelCase__ (snake_case__ : Optional[Any] , snake_case__ : List[str]=5 ): """simple docstring""" _snake_case : Any = ["""auto-generated""", """autogenerated""", """automatically generated"""] _snake_case : Tuple = example["""content"""].splitlines() for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Union[str, Any]=5 , snake_case__ : Any=0.05 ): """simple docstring""" _snake_case : Optional[Any] = ["""unit tests""", """test file""", """configuration file"""] _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : Dict = 0 _snake_case : str = 0 # first test for _, line in zip(range(snake_case__ ) , snake_case__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _snake_case : Optional[int] = example["""content"""].count("""\n""" ) _snake_case : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : Optional[int] = ["""def """, """class """, """for """, """while """] _snake_case : str = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ (snake_case__ : List[str] , snake_case__ : List[str]=4 ): """simple docstring""" _snake_case : List[Any] = example["""content"""].splitlines() _snake_case : str = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ (snake_case__ : List[str] ): """simple docstring""" _snake_case : Optional[Any] = tokenizer(example["""content"""] , truncation=snake_case__ )["""input_ids"""] _snake_case : Optional[Any] = len(example["""content"""] ) / len(snake_case__ ) return {"ratio": ratio} def UpperCAmelCase__ (snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = {} results.update(get_hash(snake_case__ ) ) results.update(line_stats(snake_case__ ) ) results.update(alpha_stats(snake_case__ ) ) results.update(char_token_ratio(snake_case__ ) ) results.update(is_autogenerated(snake_case__ ) ) results.update(is_config_or_test(snake_case__ ) ) results.update(has_no_keywords(snake_case__ ) ) results.update(has_few_assignments(snake_case__ ) ) return results def UpperCAmelCase__ (snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] ): """simple docstring""" if not check_uniques(snake_case__ , snake_case__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """rb""" ) as f_in: with gzip.open(str(snake_case__ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(snake_case__ , snake_case__ ) os.unlink(snake_case__ ) # Settings A_ = HfArgumentParser(PreprocessingArguments) A_ = parser.parse_args() if args.num_workers is None: A_ = multiprocessing.cpu_count() A_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ = time.time() A_ = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ = time.time() A_ = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ = set(ds.unique('''hash''')) A_ = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ = time.time() A_ = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ = time.time() A_ , A_ = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) A_ = output_dir / '''data''' data_dir.mkdir(exist_ok=True) A_ = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ = str(data_dir / F'''file-{file_number+1:012}.json''') A_ = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
64
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/config.json''', # See all XGLM models at https://huggingface.co/models?filter=xglm } class lowercase( __a ): '''simple docstring''' lowercase__ = "xglm" lowercase__ = ["past_key_values"] lowercase__ = { "num_attention_heads": "attention_heads", "hidden_size": "d_model", "num_hidden_layers": "num_layers", } def __init__( self: Dict, a_: int=256_008, a_: List[str]=2_048, a_: Dict=1_024, a_: int=4_096, a_: List[Any]=24, a_: Any=16, a_: Dict="gelu", a_: Optional[Any]=0.1, a_: str=0.1, a_: Union[str, Any]=0.0, a_: List[str]=0.0, a_: List[Any]=0.02, a_: Dict=True, a_: int=True, a_: List[Any]=2, a_: str=1, a_: Optional[int]=0, a_: Tuple=2, **a_: Tuple, ): '''simple docstring''' _snake_case : Union[str, Any] = vocab_size _snake_case : Optional[int] = max_position_embeddings _snake_case : Union[str, Any] = d_model _snake_case : Optional[int] = ffn_dim _snake_case : List[Any] = num_layers _snake_case : int = attention_heads _snake_case : int = activation_function _snake_case : List[str] = dropout _snake_case : List[Any] = attention_dropout _snake_case : Any = activation_dropout _snake_case : Union[str, Any] = layerdrop _snake_case : int = init_std _snake_case : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _snake_case : Union[str, Any] = use_cache super().__init__( pad_token_id=a_, bos_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, **a_, )
64
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
1
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class lowercase: '''simple docstring''' def __init__( self: Union[str, Any], a_: Optional[Any], a_: int, a_: int ): '''simple docstring''' if dst_width < 0 or dst_height < 0: raise ValueError("""Destination width/height should be > 0""" ) _snake_case : Dict = img _snake_case : Union[str, Any] = img.shape[1] _snake_case : int = img.shape[0] _snake_case : int = dst_width _snake_case : Tuple = dst_height _snake_case : Any = self.src_w / self.dst_w _snake_case : Union[str, Any] = self.src_h / self.dst_h _snake_case : Optional[int] = ( np.ones((self.dst_h, self.dst_w, 3), np.uinta ) * 255 ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' for i in range(self.dst_h ): for j in range(self.dst_w ): _snake_case : Dict = self.img[self.get_y(a_ )][self.get_x(a_ )] def UpperCamelCase_ ( self: List[str], a_: int ): '''simple docstring''' return int(self.ratio_x * x ) def UpperCamelCase_ ( self: Optional[Any], a_: int ): '''simple docstring''' return int(self.ratio_y * y ) if __name__ == "__main__": A_ , A_ = 8_00, 6_00 A_ = imread('''image_data/lena.jpg''', 1) A_ = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F'''Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}''', n.output ) waitKey(0) destroyAllWindows()
64
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Dict , snake_case__ : Any , snake_case__ : str , snake_case__ : str ): """simple docstring""" for attribute in key.split(""".""" ): _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ) if weight_type is not None: _snake_case : Optional[Any] = getattr(snake_case__ , snake_case__ ).shape else: _snake_case : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": _snake_case : int = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : Tuple = value elif weight_type == "bias": _snake_case : List[str] = value else: _snake_case : int = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def UpperCAmelCase__ (snake_case__ : str , snake_case__ : List[str] ): """simple docstring""" _snake_case : List[Any] = [] _snake_case : Optional[Any] = fairseq_model.state_dict() _snake_case : str = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _snake_case : Optional[Any] = None for name, value in fairseq_dict.items(): _snake_case : Optional[Any] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == """group""" , ) _snake_case : Dict = True elif name.split(""".""" )[0] == "proj": _snake_case : Dict = fairseq_model.proj _snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: _snake_case : Dict = True if "*" in mapped_key: _snake_case : Optional[int] = name.split(snake_case__ )[0].split(""".""" )[-2] _snake_case : Union[str, Any] = mapped_key.replace("""*""" , snake_case__ ) if "weight_g" in name: _snake_case : str = """weight_g""" elif "weight_v" in name: _snake_case : Optional[Any] = """weight_v""" elif "bias" in name: _snake_case : Union[str, Any] = """bias""" elif "weight" in name: _snake_case : int = """weight""" else: _snake_case : Optional[int] = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) continue if not is_used: unused_weights.append(snake_case__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def UpperCAmelCase__ (snake_case__ : Any , snake_case__ : Dict , snake_case__ : Union[str, Any] , snake_case__ : Union[str, Any] , snake_case__ : int ): """simple docstring""" _snake_case : Any = full_name.split("""conv_layers.""" )[-1] _snake_case : Optional[int] = name.split(""".""" ) _snake_case : List[str] = int(items[0] ) _snake_case : Dict = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) _snake_case : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) _snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) _snake_case : int = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) _snake_case : List[str] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" _snake_case , _snake_case : Optional[Any] = emb.weight.shape _snake_case : Optional[int] = nn.Linear(snake_case__ , snake_case__ , bias=snake_case__ ) _snake_case : Union[str, Any] = emb.weight.data return lin_layer def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Any = f.readlines() _snake_case : Optional[Any] = [line.split(""" """ )[0] for line in lines] _snake_case : str = len(snake_case__ ) _snake_case : Tuple = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(snake_case__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def UpperCAmelCase__ (snake_case__ : int , snake_case__ : List[str] , snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Union[str, Any] , ): """simple docstring""" _snake_case : Optional[int] = WavaVecaConfig.from_pretrained(snake_case__ ) _snake_case : List[str] = SpeechaTextaConfig.from_pretrained( snake_case__ , vocab_size=snake_case__ , decoder_layers=snake_case__ , do_stable_layer_norm=snake_case__ ) _snake_case : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=snake_case__ , return_attention_mask=snake_case__ , ) _snake_case , _snake_case , _snake_case : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) _snake_case : Optional[Any] = model[0].eval() # set weights for wav2vec2 encoder _snake_case : Any = WavaVecaModel(snake_case__ ) _snake_case : Optional[Any] = recursively_load_weights_wavaveca(model.encoder , snake_case__ ) _snake_case : Optional[Any] = SpeechaTextaForCausalLM(snake_case__ ) _snake_case , _snake_case : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=snake_case__ ) # set output linear layer unexpected_keys.remove("""embed_out""" ) _snake_case : Any = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) _snake_case : Any = SpeechEncoderDecoderModel(encoder=snake_case__ , decoder=snake_case__ ) _snake_case : Any = False # add projection layer _snake_case : int = nn.Parameter(projection_layer.weight ) _snake_case : Any = nn.Parameter(projection_layer.bias ) _snake_case : Any = create_vocab_dict(snake_case__ ) with open(os.path.join(snake_case__ , """vocab.json""" ) , """w""" ) as fp: json.dump(snake_case__ , snake_case__ ) _snake_case : Dict = SpeechaTextaTokenizer(os.path.join(snake_case__ , """vocab.json""" ) ) tokenizer.save_pretrained(snake_case__ ) _snake_case : str = hf_wavavec.config.to_dict() _snake_case : List[str] = tokenizer.pad_token_id _snake_case : Union[str, Any] = tokenizer.bos_token_id _snake_case : Union[str, Any] = tokenizer.eos_token_id _snake_case : Optional[Any] = """speech_to_text_2""" _snake_case : Optional[int] = """wav2vec2""" _snake_case : Tuple = SpeechEncoderDecoderConfig.from_dict(snake_case__ ) hf_wavavec.save_pretrained(snake_case__ ) feature_extractor.save_pretrained(snake_case__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_02_24, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') A_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
64
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Dict = len(snake_case__ ) _snake_case : Union[str, Any] = len(snake_case__ ) _snake_case : List[str] = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _snake_case : str = True for i in range(snake_case__ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _snake_case : Optional[Any] = True if a[i].islower(): _snake_case : Union[str, Any] = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
64
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
64
"""simple docstring""" import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ (snake_case__ : Optional[int] , snake_case__ : Any=7 ): """simple docstring""" _snake_case : Any = None if token is not None: _snake_case : Any = {"""Accept""": """application/vnd.github+json""", """Authorization""": F"Bearer {token}"} # The id of a workflow (not of a workflow run) _snake_case : List[str] = """636036""" _snake_case : Union[str, Any] = F"https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs" # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F"?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}" _snake_case : str = requests.get(snake_case__ , headers=snake_case__ ).json() return result["workflow_runs"] def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" _snake_case : str = get_daily_ci_runs(snake_case__ ) _snake_case : str = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": _snake_case : List[str] = workflow_run["""id"""] break return workflow_run_id def UpperCAmelCase__ (snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Optional[Any] = get_last_daily_ci_runs(snake_case__ ) if workflow_run_id is not None: _snake_case : Optional[Any] = get_artifacts_links(worflow_run_id=snake_case__ , token=snake_case__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: _snake_case : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=snake_case__ , artifact_url=snake_case__ , output_dir=snake_case__ , token=snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int ): """simple docstring""" get_last_daily_ci_artifacts(snake_case__ , snake_case__ , snake_case__ ) _snake_case : int = {} for artifact_name in artifact_names: _snake_case : int = os.path.join(snake_case__ , F"{artifact_name}.zip" ) if os.path.isfile(snake_case__ ): _snake_case : Tuple = {} with zipfile.ZipFile(snake_case__ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case__ ): # read the file with z.open(snake_case__ ) as f: _snake_case : Any = f.read().decode("""UTF-8""" ) return results
64
1
"""simple docstring""" import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path A_ = [ {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.de'''}, {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.en'''}, {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.fr'''}, {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.frr'''}, {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.it'''}, {'''dataset''': '''wikipedia''', '''config_name''': '''20220301.simple'''}, {'''dataset''': '''snli''', '''config_name''': '''plain_text'''}, {'''dataset''': '''eli5''', '''config_name''': '''LFQA_reddit'''}, {'''dataset''': '''wiki40b''', '''config_name''': '''en'''}, {'''dataset''': '''wiki_dpr''', '''config_name''': '''psgs_w100.nq.compressed'''}, {'''dataset''': '''wiki_dpr''', '''config_name''': '''psgs_w100.nq.no_index'''}, {'''dataset''': '''wiki_dpr''', '''config_name''': '''psgs_w100.multiset.no_index'''}, {'''dataset''': '''natural_questions''', '''config_name''': '''default'''}, ] def UpperCAmelCase__ (snake_case__ : Any=True ): """simple docstring""" if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class lowercase( __a ): '''simple docstring''' lowercase__ = None lowercase__ = None def UpperCamelCase_ ( self: str, a_: Dict, a_: List[Any] ): '''simple docstring''' with TemporaryDirectory() as tmp_dir: _snake_case : Optional[int] = dataset_module_factory(a_, cache_dir=a_ ) _snake_case : Optional[Any] = import_main_class(dataset_module.module_path, dataset=a_ ) _snake_case : DatasetBuilder = builder_cls( cache_dir=a_, config_name=a_, hash=dataset_module.hash, ) _snake_case : int = """/""".join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=a_ ).replace(os.sep, """/""" ), config.DATASET_INFO_FILENAME, ] ) _snake_case : Optional[Any] = cached_path(a_, cache_dir=a_ ) self.assertTrue(os.path.exists(a_ ) ) @pytest.mark.integration def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" _snake_case : Tuple = tmp_path_factory.mktemp("""test_hf_gcp""" ) / """test_wikipedia_simple""" _snake_case : Any = dataset_module_factory("""wikipedia""" , cache_dir=snake_case__ ) _snake_case : Tuple = import_main_class(dataset_module.module_path ) _snake_case : DatasetBuilder = builder_cls( cache_dir=snake_case__ , config_name="""20220301.frr""" , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam _snake_case : Optional[Any] = None builder_instance.download_and_prepare() _snake_case : Tuple = builder_instance.as_dataset() assert ds @pytest.mark.integration def UpperCAmelCase__ (snake_case__ : Dict ): """simple docstring""" _snake_case : int = dataset_module_factory("""wikipedia""" , cache_dir=snake_case__ ) _snake_case : Tuple = import_main_class(dataset_module.module_path , dataset=snake_case__ ) _snake_case : DatasetBuilder = builder_cls( cache_dir=snake_case__ , config_name="""20220301.frr""" , hash=dataset_module.hash , ) _snake_case : str = builder_instance.as_streaming_dataset() assert ds assert isinstance(snake_case__ , snake_case__ ) assert "train" in ds assert isinstance(ds["""train"""] , snake_case__ ) assert next(iter(ds["""train"""] ) )
64
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging A_ = logging.get_logger(__name__) class lowercase: '''simple docstring''' lowercase__ = 42 lowercase__ = None @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Tuple, a_: int, a_: int, a_: str, **a_: Dict ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any], a_: List[str] ): '''simple docstring''' raise NotImplementedError def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' if not self.is_available(): raise RuntimeError( f"You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}." ) @classmethod def UpperCamelCase_ ( cls: Tuple ): '''simple docstring''' return f"`pip install {cls.pip_package or cls.name}`" class lowercase( __a ): '''simple docstring''' lowercase__ = "optuna" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_optuna_available() def UpperCamelCase_ ( self: Union[str, Any], a_: List[Any], a_: int, a_: str, **a_: List[str] ): '''simple docstring''' return run_hp_search_optuna(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], a_: Any ): '''simple docstring''' return default_hp_space_optuna(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "ray" lowercase__ = "'ray[tune]'" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_ray_available() def UpperCamelCase_ ( self: int, a_: Optional[Any], a_: int, a_: str, **a_: List[Any] ): '''simple docstring''' return run_hp_search_ray(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Tuple ): '''simple docstring''' return default_hp_space_ray(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "sigopt" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_sigopt_available() def UpperCamelCase_ ( self: Dict, a_: str, a_: int, a_: str, **a_: int ): '''simple docstring''' return run_hp_search_sigopt(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: List[str] ): '''simple docstring''' return default_hp_space_sigopt(a_ ) class lowercase( __a ): '''simple docstring''' lowercase__ = "wandb" @staticmethod def UpperCamelCase_ ( ): '''simple docstring''' return is_wandb_available() def UpperCamelCase_ ( self: Optional[Any], a_: str, a_: int, a_: str, **a_: Union[str, Any] ): '''simple docstring''' return run_hp_search_wandb(a_, a_, a_, **a_ ) def UpperCamelCase_ ( self: str, a_: Any ): '''simple docstring''' return default_hp_space_wandb(a_ ) A_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def UpperCAmelCase__ (): """simple docstring""" _snake_case : Optional[int] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(snake_case__ ) > 0: _snake_case : Any = available_backends[0].name if len(snake_case__ ) > 1: logger.info( F"{len(snake_case__ )} hyperparameter search backends available. Using {name} as the default." ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( F" - To install {backend.name} run {backend.pip_install()}" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
64
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: A_ = None A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''moussaKam/mbarthez''': '''https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json''', '''moussaKam/barthez''': '''https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json''', '''moussaKam/barthez-orangesum-title''': ( '''https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json''' ), }, } A_ = { '''moussaKam/mbarthez''': 10_24, '''moussaKam/barthez''': 10_24, '''moussaKam/barthez-orangesum-title''': 10_24, } A_ = '''▁''' class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = BarthezTokenizer def __init__( self: Tuple, a_: Union[str, Any]=None, a_: Dict=None, a_: Tuple="<s>", a_: int="</s>", a_: Dict="</s>", a_: Dict="<s>", a_: Any="<unk>", a_: Tuple="<pad>", a_: Tuple="<mask>", **a_: Tuple, ): '''simple docstring''' _snake_case : List[Any] = AddedToken(a_, lstrip=a_, rstrip=a_ ) if isinstance(a_, a_ ) else mask_token super().__init__( a_, tokenizer_file=a_, bos_token=a_, eos_token=a_, unk_token=a_, sep_token=a_, cls_token=a_, pad_token=a_, mask_token=a_, **a_, ) _snake_case : Optional[int] = vocab_file _snake_case : List[str] = False if not self.vocab_file else True def UpperCamelCase_ ( self: Optional[Any], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _snake_case : Optional[int] = [self.cls_token_id] _snake_case : Dict = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase_ ( self: Tuple, a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : str = [self.sep_token_id] _snake_case : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase_ ( self: List[Any], a_: str, a_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(a_ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _snake_case : int = os.path.join( a_, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file, a_ ) return (out_vocab_file,)
64
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
1
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A_ = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = XLMRobertaTokenizer lowercase__ = XLMRobertaTokenizerFast lowercase__ = True lowercase__ = True def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _snake_case : Any = XLMRobertaTokenizer(a_, keep_accents=a_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : int = """<pad>""" _snake_case : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ), a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ), a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], """<s>""" ) self.assertEqual(vocab_keys[1], """<pad>""" ) self.assertEqual(vocab_keys[-1], """<mask>""" ) self.assertEqual(len(a_ ), 1_002 ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size, 1_002 ) def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Dict = XLMRobertaTokenizer(a_, keep_accents=a_ ) _snake_case : int = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a_, ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a_ ), [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]], ) _snake_case : Tuple = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a_, [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ], ) _snake_case : Union[str, Any] = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_, [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ], ) _snake_case : Optional[Any] = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_, [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ], ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _snake_case : Optional[int] = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : Tuple = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : int = self.tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : Union[str, Any] = tempfile.mkdtemp() _snake_case : str = tokenizer_r.save_pretrained(a_ ) _snake_case : List[Any] = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) _snake_case : List[Any] = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(a_, a_ ) # Checks everything loads correctly in the same way _snake_case : List[str] = tokenizer_r.from_pretrained(a_ ) _snake_case : Union[str, Any] = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=True _snake_case : Optional[Any] = tempfile.mkdtemp() _snake_case : Tuple = tokenizer_r.save_pretrained(a_, legacy_format=a_ ) _snake_case : str = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files self.assertSequenceEqual(a_, a_ ) # Checks everything loads correctly in the same way _snake_case : Any = tokenizer_r.from_pretrained(a_ ) _snake_case : Any = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=False _snake_case : Union[str, Any] = tempfile.mkdtemp() _snake_case : Dict = tokenizer_r.save_pretrained(a_, legacy_format=a_ ) _snake_case : List[str] = tokenizer_p.save_pretrained(a_ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _snake_case : Optional[int] = tokenizer_r.from_pretrained(a_ ) _snake_case : Optional[int] = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_, a_ ) ) shutil.rmtree(a_ ) @cached_property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(a_, f.name ) _snake_case : str = XLMRobertaTokenizer(f.name, keep_accents=a_ ) _snake_case : str = pickle.dumps(a_ ) pickle.loads(a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : int = self.get_rust_tokenizer() _snake_case : Optional[Any] = """I was born in 92000, and this is falsé.""" _snake_case : List[str] = tokenizer.tokenize(a_ ) _snake_case : str = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) _snake_case : List[str] = tokenizer.encode(a_, add_special_tokens=a_ ) _snake_case : Optional[int] = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) _snake_case : List[Any] = self.get_rust_tokenizer() _snake_case : Tuple = tokenizer.encode(a_ ) _snake_case : Dict = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) @slow def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Union[str, Any] = """Hello World!""" _snake_case : str = [0, 35_378, 6_661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(a_, self.big_tokenizer.encode(a_ ) ) @slow def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : List[Any] = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) _snake_case : Optional[int] = [ 0, 3_293, 83, 10, 4_552, 4_989, 7_986, 678, 10, 5_915, 111, 179_459, 124_850, 4, 6_044, 237, 12, 6, 5, 6, 4, 6_780, 705, 15, 1_388, 44, 378, 10_114, 711, 152, 20, 6, 5, 22_376, 642, 1_221, 15_190, 34_153, 450, 5_608, 959, 1_119, 57_702, 136, 186, 47, 1_098, 29_367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_044, 237, 6_284, 50_901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(a_, self.big_tokenizer.encode(a_ ) ) @slow def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' _snake_case : List[str] = {"""input_ids""": [[0, 11_062, 82_772, 7, 15, 82_772, 538, 51_529, 237, 17_198, 1_290, 206, 9, 215_175, 1_314, 136, 17_198, 1_290, 206, 9, 56_359, 42, 122_009, 9, 16_466, 16, 87_344, 4_537, 9, 4_717, 78_381, 6, 159_958, 7, 15, 24_480, 618, 4, 527, 22_693, 5_428, 4, 2_777, 24_480, 9_874, 4, 43_523, 594, 4, 803, 18_392, 33_189, 18, 4, 43_523, 24_447, 12_399, 100, 24_955, 83_658, 9_626, 144_057, 15, 839, 22_335, 16, 136, 24_955, 83_658, 83_479, 15, 39_102, 724, 16, 678, 645, 2_789, 1_328, 4_589, 42, 122_009, 115_774, 23, 805, 1_328, 46_876, 7, 136, 53_894, 1_940, 42_227, 41_159, 17_721, 823, 425, 4, 27_512, 98_722, 206, 136, 5_531, 4_970, 919, 17_336, 5, 2], [0, 20_080, 618, 83, 82_775, 47, 479, 9, 1_517, 73, 53_894, 333, 80_581, 110_117, 18_811, 5_256, 1_295, 51, 152_526, 297, 7_986, 390, 124_416, 538, 35_431, 214, 98, 15_044, 25_737, 136, 7_108, 43_701, 23, 756, 135_355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63_773, 119_455, 6, 147_797, 88_203, 7, 645, 70, 21, 3_285, 10_269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_, model_name="""xlm-roberta-base""", revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""", )
64
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : int = 0.00 _snake_case : int = 0 for resistor in resistors: if resistor <= 0: _snake_case : Dict = F"Resistor at index {index} has a negative or zero value!" raise ValueError(snake_case__ ) first_sum += 1 / float(snake_case__ ) index += 1 return 1 / first_sum def UpperCAmelCase__ (snake_case__ : list[float] ): """simple docstring""" _snake_case : Union[str, Any] = 0.00 _snake_case : Any = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _snake_case : Any = F"Resistor at index {index} has a negative value!" raise ValueError(snake_case__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
64
1
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging A_ = logging.get_logger(__name__) class lowercase( __a ): '''simple docstring''' lowercase__ = ["pixel_values"] def __init__( self: Dict, a_: bool = True, a_: Union[int, float] = 1 / 255, a_: bool = True, a_: int = 8, **a_: str, ): '''simple docstring''' super().__init__(**a_ ) _snake_case : List[Any] = do_rescale _snake_case : Tuple = rescale_factor _snake_case : Dict = do_pad _snake_case : List[Any] = pad_size def UpperCamelCase_ ( self: List[Any], a_: np.ndarray, a_: float, a_: Optional[Union[str, ChannelDimension]] = None, **a_: Tuple ): '''simple docstring''' return rescale(a_, scale=a_, data_format=a_, **a_ ) def UpperCamelCase_ ( self: str, a_: np.ndarray, a_: int, a_: Optional[Union[str, ChannelDimension]] = None ): '''simple docstring''' _snake_case , _snake_case : List[str] = get_image_size(a_ ) _snake_case : Union[str, Any] = (old_height // size + 1) * size - old_height _snake_case : str = (old_width // size + 1) * size - old_width return pad(a_, ((0, pad_height), (0, pad_width)), mode="""symmetric""", data_format=a_ ) def UpperCamelCase_ ( self: int, a_: ImageInput, a_: Optional[bool] = None, a_: Optional[float] = None, a_: Optional[bool] = None, a_: Optional[int] = None, a_: Optional[Union[str, TensorType]] = None, a_: Union[str, ChannelDimension] = ChannelDimension.FIRST, **a_: Union[str, Any], ): '''simple docstring''' _snake_case : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale _snake_case : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case : List[str] = do_pad if do_pad is not None else self.do_pad _snake_case : Optional[int] = pad_size if pad_size is not None else self.pad_size _snake_case : Any = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. _snake_case : int = [to_numpy_array(a_ ) for image in images] if do_rescale: _snake_case : Dict = [self.rescale(image=a_, scale=a_ ) for image in images] if do_pad: _snake_case : Union[str, Any] = [self.pad(a_, size=a_ ) for image in images] _snake_case : Tuple = [to_channel_dimension_format(a_, a_ ) for image in images] _snake_case : Any = {"""pixel_values""": images} return BatchFeature(data=a_, tensor_type=a_ )
64
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 20_48, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["input_ids", "attention_mask"] lowercase__ = CodeGenTokenizer def __init__( self: Union[str, Any], a_: List[Any]=None, a_: str=None, a_: str=None, a_: Dict="<|endoftext|>", a_: Tuple="<|endoftext|>", a_: str="<|endoftext|>", a_: List[Any]=False, **a_: List[str], ): '''simple docstring''' super().__init__( a_, a_, tokenizer_file=a_, unk_token=a_, bos_token=a_, eos_token=a_, add_prefix_space=a_, **a_, ) if kwargs.pop("""add_bos_token""", a_ ): _snake_case : str = kwargs.pop("""name_or_path""", """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" f"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" f"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) _snake_case : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""", a_ ) != add_prefix_space: _snake_case : Dict = getattr(a_, pre_tok_state.pop("""type""" ) ) _snake_case : Dict = add_prefix_space _snake_case : str = pre_tok_class(**a_ ) _snake_case : List[Any] = add_prefix_space def UpperCamelCase_ ( self: Any, *a_: Any, **a_: int ): '''simple docstring''' _snake_case : Optional[int] = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[Any], *a_: Any, **a_: List[str] ): '''simple docstring''' _snake_case : Dict = kwargs.get("""is_split_into_words""", a_ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*a_, **a_ ) def UpperCamelCase_ ( self: Optional[int], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ ) def UpperCamelCase_ ( self: str, a_: Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"], a_: bool = False, a_: bool = None, a_: Optional[List[str]] = None, **a_: List[str], ): '''simple docstring''' _snake_case : Any = super().decode( token_ids=a_, skip_special_tokens=a_, clean_up_tokenization_spaces=a_, **a_, ) if truncate_before_pattern is not None and len(a_ ) > 0: _snake_case : List[str] = self.truncate(a_, a_ ) return decoded_text def UpperCamelCase_ ( self: Dict, a_: Tuple, a_: Optional[Any] ): '''simple docstring''' def find_re(a_: Dict, a_: str, a_: Union[str, Any] ): _snake_case : Any = pattern.search(a_, a_ ) return m.start() if m else -1 _snake_case : Tuple = [re.compile(a_, re.MULTILINE ) for pattern in truncate_before_pattern] _snake_case : List[Any] = list(re.finditer("""^print""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : int = completion[: prints[1].start()] _snake_case : List[str] = list(re.finditer("""^def""", a_, re.MULTILINE ) ) if len(a_ ) > 1: _snake_case : List[Any] = completion[: defs[1].start()] _snake_case : int = 0 _snake_case : List[Any] = [ pos for pos in [find_re(a_, a_, a_ ) for terminal in terminals] if pos != -1 ] if len(a_ ) > 0: return completion[: min(a_ )] else: return completion
64
1
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: List[str], a_: List[str]=None, a_: Tuple=None, **a_: Tuple ): '''simple docstring''' _snake_case : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : str = kwargs.pop("""feature_extractor""" ) _snake_case : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Any = False def __call__( self: Any, *a_: Any, **a_: Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : Dict = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Optional[int] = args[0] _snake_case : Tuple = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : Tuple = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : Tuple = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : List[str] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: Optional[int], *a_: Tuple, **a_: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: int, *a_: List[str], **a_: int ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : Any = True _snake_case : Optional[int] = self.tokenizer yield _snake_case : int = self.image_processor _snake_case : Optional[int] = False def UpperCamelCase_ ( self: Dict, a_: Optional[Any], a_: str=False, a_: Optional[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : Dict = self.tokenizer.get_added_vocab() _snake_case : str = {} while tokens: _snake_case : Union[str, Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : List[Any] = start_token.group(1 ) _snake_case : str = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Dict = start_token.group() if end_token is None: _snake_case : List[Any] = tokens.replace(a_, """""" ) else: _snake_case : List[str] = end_token.group() _snake_case : str = re.escape(a_ ) _snake_case : str = re.escape(a_ ) _snake_case : Union[str, Any] = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : List[Any] = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : List[str] = value[0] _snake_case : List[str] = value else: # leaf nodes _snake_case : Tuple = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : Tuple = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : int = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Any = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
64
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" _snake_case : List[Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _snake_case : Tuple = 1_92 _snake_case : Any = 7_68 _snake_case : Any = 12 _snake_case : List[Any] = 3 _snake_case : int = [8_00, 13_33] _snake_case : Tuple = False elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = 3_30 _snake_case : List[str] = 14 _snake_case : List[str] = 6 _snake_case : Union[str, Any] = 13_20 elif "yolos_s" in yolos_name: _snake_case : Union[str, Any] = 3_84 _snake_case : List[str] = 15_36 _snake_case : Any = 12 _snake_case : Optional[int] = 6 elif "yolos_b" in yolos_name: _snake_case : Dict = [8_00, 13_44] _snake_case : str = 91 _snake_case : Optional[Any] = """huggingface/label-files""" _snake_case : str = """coco-detection-id2label.json""" _snake_case : str = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Union[str, Any] = {int(snake_case__ ): v for k, v in idalabel.items()} _snake_case : List[str] = idalabel _snake_case : List[str] = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosConfig , snake_case__ : bool = False ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) _snake_case : Union[str, Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict _snake_case : Any = in_proj_weight[: config.hidden_size, :] _snake_case : Optional[Any] = in_proj_bias[: config.hidden_size] _snake_case : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Tuple = in_proj_weight[-config.hidden_size :, :] _snake_case : List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ (snake_case__ : str ): """simple docstring""" if "backbone" in name: _snake_case : str = name.replace("""backbone""" , """vit""" ) if "cls_token" in name: _snake_case : Union[str, Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "det_token" in name: _snake_case : str = name.replace("""det_token""" , """embeddings.detection_tokens""" ) if "mid_pos_embed" in name: _snake_case : str = name.replace("""mid_pos_embed""" , """encoder.mid_position_embeddings""" ) if "pos_embed" in name: _snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: _snake_case : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "blocks" in name: _snake_case : str = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: _snake_case : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _snake_case : str = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _snake_case : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _snake_case : str = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _snake_case : int = name.replace("""mlp.fc2""" , """output.dense""" ) if "class_embed" in name: _snake_case : Union[str, Any] = name.replace("""class_embed""" , """class_labels_classifier""" ) if "bbox_embed" in name: _snake_case : str = name.replace("""bbox_embed""" , """bbox_predictor""" ) if "vit.norm" in name: _snake_case : Union[str, Any] = name.replace("""vit.norm""" , """vit.layernorm""" ) return name def UpperCAmelCase__ (snake_case__ : dict , snake_case__ : YolosForObjectDetection ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[str] = orig_state_dict.pop(snake_case__ ) if "qkv" in key: _snake_case : Optional[Any] = key.split(""".""" ) _snake_case : Optional[Any] = int(key_split[2] ) _snake_case : Optional[int] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _snake_case : str = val[:dim, :] _snake_case : Optional[Any] = val[ dim : dim * 2, : ] _snake_case : Optional[Any] = val[-dim:, :] else: _snake_case : Dict = val[:dim] _snake_case : Any = val[dim : dim * 2] _snake_case : Dict = val[-dim:] else: _snake_case : Tuple = val return orig_state_dict def UpperCAmelCase__ (): """simple docstring""" _snake_case : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : str , snake_case__ : bool = False ): """simple docstring""" _snake_case : Optional[Any] = get_yolos_config(snake_case__ ) # load original state_dict _snake_case : Optional[int] = torch.load(snake_case__ , map_location="""cpu""" )["""model"""] # load 🤗 model _snake_case : Optional[Any] = YolosForObjectDetection(snake_case__ ) model.eval() _snake_case : Optional[Any] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor _snake_case : List[str] = 8_00 if yolos_name != """yolos_ti""" else 5_12 _snake_case : Optional[int] = YolosImageProcessor(format="""coco_detection""" , size=snake_case__ ) _snake_case : Optional[Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _snake_case : Optional[Any] = model(**snake_case__ ) _snake_case , _snake_case : Optional[int] = outputs.logits, outputs.pred_boxes _snake_case , _snake_case : Dict = None, None if yolos_name == "yolos_ti": _snake_case : Optional[Any] = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]] ) _snake_case : Tuple = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]] ) elif yolos_name == "yolos_s_200_pre": _snake_case : List[str] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] ) _snake_case : List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] ) elif yolos_name == "yolos_s_300_pre": _snake_case : Dict = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]] ) _snake_case : Union[str, Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]] ) elif yolos_name == "yolos_s_dWr": _snake_case : Tuple = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]] ) _snake_case : Optional[Any] = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]] ) elif yolos_name == "yolos_base": _snake_case : int = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]] ) _snake_case : Optional[int] = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: _snake_case : Dict = { """yolos_ti""": """yolos-tiny""", """yolos_s_200_pre""": """yolos-small""", """yolos_s_300_pre""": """yolos-small-300""", """yolos_s_dWr""": """yolos-small-dwr""", """yolos_base""": """yolos-base""", } print("""Pushing to the hub...""" ) _snake_case : str = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="""hustvl""" ) model.push_to_hub(snake_case__ , organization="""hustvl""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
64
1